messages
listlengths 2
2
|
|---|
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: src1: Vale.X64.Machine_s.quad32 -> src2: Vale.X64.Machine_s.quad32\n -> FStar.Pervasives.Native.option Vale.X64.Machine_s.quad32\nEffect: Prims.Tot\n\nContext:\nmodule Vale.X64.Instructions_s\nopen FStar.Mul\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Two_s\nopen Vale.Def.Words.Four_s\nopen Vale.Def.Types_s\nopen Vale.X64.Machine_s\nopen Vale.X64.Instruction_s\nopen Vale.X64.CPU_Features_s\nopen Vale.X64.CryptoInstructions_s\nopen FStar.Seq.Base\n\nlet eval_Mov64 (src:nat64) : option nat64 = Some src\nval ins_Mov64 : instr_dep [out op64] [op64] PreserveFlags eval_Mov64\n\nlet eval_MovBe64 (src:nat64) : option nat64 = if movbe_enabled then Some (reverse_bytes_nat64 src) else None\nval ins_MovBe64 : instr_dep [out op64] [op64] PreserveFlags eval_MovBe64\n\nlet eval_Bswap64 (dst:nat64) : option nat64 = Some (reverse_bytes_nat64 dst)\nval ins_Bswap64 : instr_dep [inOut op64] [] PreserveFlags eval_Bswap64\n\nlet eval_Cmovc64 (dst src:nat64) (carry:bool) : option nat64 = Some (if carry then src else dst)\nval ins_Cmovc64 : instr_dep [inOut op64] [op64; opFlagsCf] PreserveFlags eval_Cmovc64\n\nlet eval_Add64 (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src in Some (sum >= pow2_64, sum % pow2_64)\nval ins_Add64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Add64\n\nlet eval_AddLea64 (src1 src2:nat64) : option nat64 = Some ((src1 + src2) % pow2_64)\nval ins_AddLea64 :instr_dep [out op64] [op64; op64] PreserveFlags eval_AddLea64\n\nlet eval_AddCarry64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_carry then 1 else 0) in\n Some (sum >= pow2_64, sum % pow2_64)\nval ins_AddCarry64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_AddCarry64\n\nlet eval_Adcx64_Adox64 (old_flag:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_flag then 1 else 0) in\n if adx_enabled then Some (sum >= pow2_64, sum % pow2_64) else None\nval ins_Adcx64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\nval ins_Adox64 : instr_dep [inOut opFlagsOf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\n\nlet eval_Sub64 (dst src:nat64) : option (bool & nat64) =\n let diff = dst - src in Some (diff < 0, diff % pow2_64)\nval ins_Sub64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Sub64\n\nlet eval_Sbb64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let diff = dst - (src + (if old_carry then 1 else 0)) in\n Some (diff < 0, diff % pow2_64) // We specify cf, but underspecify everything else (via HavocFlags)\nval ins_Sbb64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_Sbb64\n\nlet eval_Mul64 (rax src:nat64) : option (nat64 & nat64) =\n Some (FStar.UInt.mul_div #64 rax src, FStar.UInt.mul_mod #64 rax src)\nval ins_Mul64 : instr_dep [out (one64Reg rRdx); inOut (one64Reg rRax)] [op64] HavocFlags eval_Mul64\n\nlet eval_Mulx64 (rdx src:nat64) : option (nat64 & nat64) =\n let hi = FStar.UInt.mul_div #64 rdx src in\n let lo = FStar.UInt.mul_mod #64 rdx src in\n if bmi2_enabled then Some (hi, lo) else None\nval ins_Mulx64 : instr_dep [out op64; out op64] [one64Reg rRdx; op64] PreserveFlags eval_Mulx64\n\nlet eval_IMul64 (dst src:nat64) : option nat64 =\n Some (FStar.UInt.mul_mod #64 dst src)\nval ins_IMul64 : instr_dep [inOut op64] [op64] HavocFlags eval_IMul64\n\nlet eval_And64 (dst src:nat64) : option nat64 = Some (iand dst src)\nval ins_And64 : instr_dep [inOut op64] [op64] HavocFlags eval_And64\n\nlet eval_Xor64 (dst src:nat64) : option (nat64 & (bool & bool)) =\n Some (Vale.Def.Types_s.ixor dst src, (false, false))\nval ins_Xor64 : instr_dep [inOut op64; out opFlagsCf; out opFlagsOf] [op64] HavocFlags eval_Xor64\n\nlet eval_Shr64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishr dst amt) else None\nval ins_Shr64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shr64\n\nlet eval_Shl64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishl dst amt) else None\nval ins_Shl64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shl64\n\nlet eval_Cpuid (rax rcx:nat64) : option (nat64 & (nat64 & (nat64 & nat64))) =\n Some (cpuid rRax rax rcx, (cpuid rRbx rax rcx, (cpuid rRcx rax rcx, cpuid rRdx rax rcx)))\nval ins_Cpuid :\n instr_dep\n [inOut (one64Reg rRax); out (one64Reg rRbx); inOut (one64Reg rRcx); out (one64Reg rRdx)]\n [] PreserveFlags eval_Cpuid\n\n// The XGETBV instruction requires that OSXSAVE (in CPUID) is enabled.\n// We underspecify XGETBV here to only support fetching XCR0, which\n// is supported on any processor supporting the XGETBV instruction\nlet eval_Xgetbv (rcx:nat64) : option (nat64 & nat64) =\n if osxsave_enabled && rcx = 0 then Some (xgetbv rRax rcx, xgetbv rRdx rcx) else None\nval ins_Xgetbv :\n instr_dep\n [out (one64Reg rRax); out (one64Reg rRdx)]\n [one64Reg rRcx] PreserveFlags eval_Xgetbv\n\nlet check_avx (#a:Type0) (x:option a) : option a =\n if avx_enabled then x else None\n\nlet check_sse2 (#a:Type0) (x:option a) : option a =\n if sse2_enabled then x else None\n\nlet check_ssse3 (#a:Type0) (x:option a) : option a =\n if ssse3_enabled then x else None\n\nlet check_sse4_1 (#a:Type0) (x:option a) : option a =\n if sse4_1_enabled then x else None\n\nlet eval_Movdqu (src:quad32) : option quad32 = check_sse2 (Some src)\nval ins_Movdqu : instr_dep [out opXmm] [opXmm] PreserveFlags eval_Movdqu\n\nlet eval_Pxor (dst src:quad32) : option quad32 = check_sse2 (Some (quad32_xor dst src))\nval ins_Pxor : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pxor\n\nlet eval_VPxor (src1 src2:quad32) : option quad32 = check_avx (Some (quad32_xor src1 src2))\nval ins_VPxor : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPxor\n\nlet eval_Pand (dst src:quad32) : option quad32 = check_sse2 (Some (four_map2 (fun di si -> iand di si) dst src))\nval ins_Pand : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pand\n\nlet eval_Paddd_raw (src1 src2:quad32) : option quad32 =\n Some (Mkfour\n ((src1.lo0 + src2.lo0) % pow2_32)\n ((src1.lo1 + src2.lo1) % pow2_32)\n ((src1.hi2 + src2.hi2) % pow2_32)\n ((src1.hi3 + src2.hi3) % pow2_32))\n\nlet eval_Paddd (src1 src2:quad32) : option quad32 = check_sse2 (eval_Paddd_raw src1 src2)\nval ins_Paddd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Paddd\n\nlet eval_VPaddd (src1 src2:quad32) : option quad32 = check_avx (eval_Paddd_raw src1 src2)\nval ins_VPaddd : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPaddd\n\nlet eval_Pslld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishl i amt) dst) else None)\nval ins_Pslld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Pslld amt)\n\nlet eval_Psrld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishr i amt) dst) else None)\nval ins_Psrld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrld amt)\n\nlet eval_Psrldq (amt:int) (dst:quad32) : option quad32 = check_sse2 (\n if 0 <= amt && amt < 16 then\n let src_bytes = le_quad32_to_bytes dst in\n let zero_pad = Seq.create amt 0 in\n let remaining_bytes = slice src_bytes amt (length src_bytes) in\n Some (le_bytes_to_quad32 (append zero_pad remaining_bytes))\n else None)\nval ins_Psrldq (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrldq amt)\n\nlet eval_Palignr_raw (amount:nat8) (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if amount = 4 then\n Some (Mkfour src2.lo1 src2.hi2 src2.hi3 src1.lo0)\n else if amount = 8 then\n Some (Mkfour src2.hi2 src2.hi3 src1.lo0 src1.lo1)\n else None\n\nlet eval_Palignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_ssse3 (eval_Palignr_raw amount src1 src2)\nval ins_Palignr (amount:nat8) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Palignr amount)\n\nlet eval_VPalignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Palignr_raw amount src1 src2)\nval ins_VPalignr (amount:nat8) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPalignr amount)\n\nlet eval_Shufpd_raw (permutation:int) (src1 src2:quad32) : option quad32 =\n if 0 <= permutation && permutation < 4 then\n Some (Mkfour\n (if permutation % 2 = 0 then src1.lo0 else src1.hi2)\n (if permutation % 2 = 0 then src1.lo1 else src1.hi3)\n (if (permutation / 2) % 2 = 0 then src2.lo0 else src2.hi2)\n (if (permutation / 2) % 2 = 0 then src2.lo1 else src2.hi3))\n else None\n\nlet eval_Shufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_sse2 (eval_Shufpd_raw permutation src1 src2)\nval ins_Shufpd (permutation:int) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Shufpd permutation)\n\nlet eval_VShufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Shufpd_raw permutation src1 src2)\nval ins_VShufpd (permutation:int) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VShufpd permutation)\n\nlet is_full_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x04050607 &&\n q.hi3 = 0x00010203\n\nlet is_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x00010203 &&\n q.lo1 = 0x04050607 &&\n q.hi2 = 0x08090A0B &&\n q.hi3 = 0x0C0D0E0F\n\nlet is_high_dup_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet is_lower_upper_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x04050607 &&\n q.lo1 = 0x00010203 &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet eval_Pshufb_raw (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if is_full_byte_reversal_mask src2 then\n Some (reverse_bytes_quad32 src1)\n else if is_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3))\n else if is_high_dup_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else if is_lower_upper_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else None\n\nlet eval_Pshufb (src1 src2:quad32) : option quad32 = check_ssse3 (eval_Pshufb_raw src1 src2)\nval ins_Pshufb : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pshufb\n\nlet eval_VPshufb (src1 src2:quad32) : option quad32 = check_avx (eval_Pshufb_raw src1 src2)\nval ins_VPshufb : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPshufb\n\nlet eval_Pshufd (permutation:nat8) (src:quad32) : option quad32 = check_sse2 (\n let bits:bits_of_byte = byte_to_twobits permutation in\n Some (Mkfour\n (select_word src bits.lo0)\n (select_word src bits.lo1)\n (select_word src bits.hi2)\n (select_word src bits.hi3)))\nval ins_Pshufd (permutation:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_Pshufd permutation)\n\nlet eval_Pcmpeqd (dst src:quad32) : option quad32 = check_sse2 (\n let eq_result (b:bool):nat32 = if b then 0xFFFFFFFF else 0 in\n Some (Mkfour\n (eq_result (src.lo0 = dst.lo0))\n (eq_result (src.lo1 = dst.lo1))\n (eq_result (src.hi2 = dst.hi2))\n (eq_result (src.hi3 = dst.hi3))))\nval ins_Pcmpeqd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pcmpeqd\n\nlet eval_Pextrq (index:nat8) (src:quad32) : option nat64 = check_sse4_1 (\n let src_two = four_to_two_two src in\n Some (two_to_nat 32 (two_select src_two (index % 2))))\nval ins_Pextrq (index:nat8) : instr_dep [out op64] [opXmm] PreserveFlags (eval_Pextrq index)\n\nlet eval_Pinsrd (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat32 dst (src % pow2_32) (index % 4)))\nval ins_Pinsrd (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrd index)\n\nlet eval_Pinsrq (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat64_def dst src (index % 2)))\nval ins_Pinsrq (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrq index)\n\nlet eval_Pslldq_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the two very special cases we need\n if count = 4 then Some (Mkfour 0 src.lo0 src.lo1 src.hi2)\n else if count = 8 then Some (Mkfour 0 0 src.lo0 src.lo1)\n else None\nlet eval_VPslldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Pslldq_raw count src)\nval ins_VPslldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPslldq count)\n\nlet eval_Psrldq_8_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the one very special case we need\n if count = 8 then Some (Mkfour src.hi2 src.hi3 0 0)\n else None\nlet eval_VPsrldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Psrldq_8_raw count src)\nval ins_VPsrldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPsrldq count)\n\nlet eval_Pclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n let Mkfour a0 a1 a2 a3 = src1 in\n let Mkfour b0 b1 b2 b3 = src2 in\n let f x0 x1 y0 y1 =\n let x = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo x0 x1) in\n let y = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo y0 y1) in\n Vale.Math.Poly2.Bits_s.to_quad32 (Vale.Math.Poly2_s.mul x y)\n in\n if pclmulqdq_enabled then\n match imm with\n | 0 -> Some (f a0 a1 b0 b1)\n | 1 -> Some (f a2 a3 b0 b1)\n | 16 -> Some (f a0 a1 b2 b3)\n | 17 -> Some (f a2 a3 b2 b3)\n | _ -> None\n else None\nval ins_Pclmulqdq (imm:int) : instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Pclmulqdq imm)\n\nlet eval_VPclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Pclmulqdq imm src1 src2)\nval ins_VPclmulqdq (imm:int) : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPclmulqdq imm)\n\nPartial implementation:\nlet eval_AESNI_enc (src1 src2: quad32) : option quad32 =\n \n"
},
{
"role": "assistant",
"content": "if aesni_enabled\nthen\n Some\n (quad32_xor (Vale.AES.AES_s.mix_columns_LE (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE\n src1)))\n src2)\nelse None"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n alg: Vale.AES.AES_common_s.algorithm ->\n in_b: Vale.X64.Memory.buffer128 ->\n out_b: Vale.X64.Memory.buffer128 ->\n key: FStar.Seq.Base.seq Vale.X64.Memory.nat32 ->\n round_keys: FStar.Seq.Base.seq Vale.X64.Decls.quad32 ->\n keys_b: Vale.X64.Memory.buffer128 ->\n hkeys_b: Vale.X64.Memory.buffer128 ->\n h_LE: Vale.X64.Decls.quad32 ->\n va_s0: Vale.X64.Decls.va_state ->\n va_k: (_: Vale.X64.Decls.va_state -> _: Prims.unit -> Type0)\n -> Prims.Ghost ((Vale.X64.Decls.va_state * Vale.X64.Decls.va_fuel) * Prims.unit)\nEffect: Prims.Ghost\n\nContext:\nmodule Vale.AES.X64.GCMdecryptOpt\nopen Vale.Def.Prop_s\nopen Vale.Def.Opaque_s\nopen FStar.Seq\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Seq_s\nopen Vale.Def.Types_s\nopen Vale.Arch.Types\nopen Vale.Arch.HeapImpl\nopen Vale.AES.AES_s\nopen Vale.AES.GCTR_s\nopen Vale.AES.GCTR\nopen Vale.AES.GCM\nopen Vale.AES.GHash_s\nopen Vale.AES.GHash\nopen Vale.AES.GCM_s\nopen Vale.AES.X64.AES\nopen Vale.AES.GF128_s\nopen Vale.AES.GF128\nopen Vale.Poly1305.Math\nopen Vale.AES.GCM_helpers\nopen Vale.AES.X64.GHash\nopen Vale.AES.X64.GCTR\nopen Vale.X64.Machine_s\nopen Vale.X64.Memory\nopen Vale.X64.Stack_i\nopen Vale.X64.State\nopen Vale.X64.Decls\nopen Vale.X64.InsBasic\nopen Vale.X64.InsMem\nopen Vale.X64.InsVector\nopen Vale.X64.InsStack\nopen Vale.X64.InsAes\nopen Vale.X64.QuickCode\nopen Vale.X64.QuickCodes\nopen Vale.AES.X64.GF128_Mul\nopen Vale.X64.Stack\nopen Vale.X64.CPU_Features_s\nopen Vale.Math.Poly2.Bits_s\nopen Vale.AES.X64.AESopt\nopen Vale.AES.X64.AESGCM\nopen Vale.AES.X64.AESopt2\nopen Vale.Lib.Meta\nopen Vale.AES.X64.GCMencryptOpt\nopen Vale.AES.OptPublic\nopen Vale.Lib.Basic\n#reset-options \"--z3rlimit 20 --max_ifuel 0\"\n//-- Gcm_extra_bytes\n\nval va_code_Gcm_extra_bytes : alg:algorithm -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_extra_bytes alg =\n (va_Block (va_CCons (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0)\n (va_op_reg_opr64_reg64 rRax) 0 Secret) (va_CCons (va_code_Mov128 (va_op_xmm_xmm 10)\n (va_op_xmm_xmm 0)) (va_CCons (va_code_Ghash_extra_bytes ()) (va_CCons (va_code_Mov128\n (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_CCons (va_code_Pshufb (va_op_xmm_xmm 0)\n (va_op_xmm_xmm 9)) (va_CCons (va_code_AESEncryptBlock alg) (va_CCons (va_code_Pxor\n (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (va_CCons (va_code_Store128_buffer\n (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax) (va_op_xmm_xmm 10) 0 Secret)\n (va_CNil ()))))))))))\n\nval va_codegen_success_Gcm_extra_bytes : alg:algorithm -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_extra_bytes alg =\n (va_pbool_and (va_codegen_success_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0)\n (va_op_reg_opr64_reg64 rRax) 0 Secret) (va_pbool_and (va_codegen_success_Mov128 (va_op_xmm_xmm\n 10) (va_op_xmm_xmm 0)) (va_pbool_and (va_codegen_success_Ghash_extra_bytes ()) (va_pbool_and\n (va_codegen_success_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_pbool_and\n (va_codegen_success_Pshufb (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (va_pbool_and\n (va_codegen_success_AESEncryptBlock alg) (va_pbool_and (va_codegen_success_Pxor (va_op_xmm_xmm\n 10) (va_op_xmm_xmm 0)) (va_pbool_and (va_codegen_success_Store128_buffer\n (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax) (va_op_xmm_xmm 10) 0 Secret)\n (va_ttrue ())))))))))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_extra_bytes (va_mods:va_mods_t) (alg:algorithm) (inout_b:buffer128) (key:(seq\n nat32)) (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (total_bytes:nat)\n (old_hash:quad32) (completed_quads:(seq quad32)) (h_LE:quad32) : (va_quickCode unit\n (va_code_Gcm_extra_bytes alg)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in let (len:(va_int_range\n 1 1)) = 1 in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 188 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64\n rRax) 0 Secret inout_b 0) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 189 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov128 (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (fun (va_s:va_state) _ -> let\n (hash_input:quad32) = va_get_xmm 0 va_s in va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 193 column 22 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Ghash_extra_bytes hkeys_b total_bytes old_hash h_LE completed_quads) (fun\n (va_s:va_state) _ -> va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 194 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (FStar.Seq.Base.equal #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s)\n inout_b) (FStar.Seq.Base.create #quad32 1 hash_input)) (let (snap:(FStar.Seq.Base.seq\n Vale.X64.Decls.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s) inout_b in va_QSeq\n va_range1\n \"***** PRECONDITION NOT MET AT line 198 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 199 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Pshufb (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (fun (va_s:va_state) _ -> va_QBind\n va_range1\n \"***** PRECONDITION NOT MET AT line 200 column 20 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_AESEncryptBlock alg (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 11 va_s)) key\n round_keys keys_b) (fun (va_s:va_state) _ -> va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 201 column 26 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.AES_s.aes_encrypt_LE_reveal ()) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 204 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Pxor (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 205 column 20 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Store128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax)\n (va_op_xmm_xmm 10) 0 Secret inout_b 0) (fun (va_s:va_state) _ -> va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 207 column 24 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.gctr_partial_reveal ()) (va_QEmpty (()))))))))))))))\n\n\nval va_lemma_Gcm_extra_bytes : va_b0:va_code -> va_s0:va_state -> alg:algorithm ->\n inout_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128 -> total_bytes:nat -> old_hash:quad32 -> completed_quads:(seq quad32) ->\n h_LE:quad32\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_extra_bytes alg) va_s0 /\\ va_get_ok va_s0 /\\ (let\n (len:(va_int_range 1 1)) = 1 in sse_enabled /\\ Vale.X64.Decls.buffers_disjoint128 keys_b\n inout_b /\\ Vale.X64.Decls.buffers_disjoint128 hkeys_b inout_b /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) (va_get_reg64 rRax va_s0) inout_b\n len (va_get_mem_layout va_s0) Secret /\\ len == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 inout_b /\\ va_get_xmm 9 va_s0 == Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\ aes_reqs alg key round_keys\n keys_b (va_get_reg64 rR8 va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\\n pclmulqdq_enabled /\\ Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0\n va_s0) hkeys_b) (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret /\\ va_get_xmm 8 va_s0 == Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.AES.GHash.ghash_incremental0 h_LE old_hash completed_quads) /\\ FStar.Seq.Base.length\n #quad32 completed_quads == total_bytes `op_Division` 16 /\\ total_bytes < 16 `op_Multiply`\n FStar.Seq.Base.length #quad32 completed_quads + 16 /\\ va_get_reg64 rR10 va_s0 == total_bytes\n `op_Modulus` 16 /\\ total_bytes `op_Modulus` 16 =!= 0 /\\ (0 < total_bytes /\\ total_bytes < 16\n `op_Multiply` Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes) /\\ 16 `op_Multiply`\n (Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes - 1) < total_bytes)))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (let (len:(va_int_range 1 1)) = 1 in Vale.X64.Decls.modifies_buffer128 inout_b\n (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5 va_sM) /\\ Vale.AES.GCTR.gctr_partial alg len\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b) key (va_get_xmm 11 va_s0) /\\ (let raw_quads =\n FStar.Seq.Base.append #quad32 completed_quads (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0)\n inout_b) in let input_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads) 0 total_bytes in let padded_bytes =\n Vale.AES.GCTR_s.pad_to_128_bits input_bytes in let input_quads =\n Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in l_and (FStar.Seq.Base.length\n #Vale.Def.Types_s.quad32 input_quads > 0) (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8\n va_sM) == Vale.AES.GHash.ghash_incremental h_LE old_hash input_quads))) /\\ va_state_eq va_sM\n (va_update_flags va_sM (va_update_mem_heaplet 5 va_sM (va_update_xmm 10 va_sM (va_update_xmm 8\n va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5 va_sM (va_update_xmm 4\n va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1 va_sM (va_update_xmm 0\n va_sM (va_update_reg64 rR11 va_sM (va_update_reg64 rRcx va_sM (va_update_ok va_sM\n (va_update_mem va_sM va_s0))))))))))))))))))\n[@\"opaque_to_smt\"]\nlet va_lemma_Gcm_extra_bytes va_b0 va_s0 alg inout_b key round_keys keys_b hkeys_b total_bytes\n old_hash completed_quads h_LE =\n let (va_mods:va_mods_t) = [va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8;\n va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm\n 1; va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_ok; va_Mod_mem] in\n let va_qc = va_qcode_Gcm_extra_bytes va_mods alg inout_b key round_keys keys_b hkeys_b\n total_bytes old_hash completed_quads h_LE in\n let (va_sM, va_fM, va_g) = va_wp_sound_code_norm (va_code_Gcm_extra_bytes alg) va_qc va_s0 (fun\n va_s0 va_sM va_g -> let () = va_g in label va_range1\n \"***** POSTCONDITION NOT MET AT line 121 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\ (let (len:(va_int_range 1 1)) = 1 in label va_range1\n \"***** POSTCONDITION NOT MET AT line 174 column 55 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 177 column 95 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.AES.GCTR.gctr_partial alg len (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM) inout_b) key (va_get_xmm 11 va_s0)) /\\ label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 180 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let raw_quads = FStar.Seq.Base.append #quad32 completed_quads (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 181 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let input_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads) 0 total_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 182 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let padded_bytes = Vale.AES.GCTR_s.pad_to_128_bits input_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 183 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let input_quads = Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 186 column 59 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (l_and (FStar.Seq.Base.length #Vale.Def.Types_s.quad32 input_quads > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE old_hash input_quads)))))))) in\n assert_norm (va_qc.mods == va_mods);\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm\n 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1;\n va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_ok; va_Mod_mem]) va_sM va_s0;\n (va_sM, va_fM)\n\n[@ va_qattr]\nlet va_wp_Gcm_extra_bytes (alg:algorithm) (inout_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (total_bytes:nat) (old_hash:quad32)\n (completed_quads:(seq quad32)) (h_LE:quad32) (va_s0:va_state) (va_k:(va_state -> unit -> Type0))\n : Type0 =\n (va_get_ok va_s0 /\\ (let (len:(va_int_range 1 1)) = 1 in sse_enabled /\\\n Vale.X64.Decls.buffers_disjoint128 keys_b inout_b /\\ Vale.X64.Decls.buffers_disjoint128 hkeys_b\n inout_b /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) (va_get_reg64 rRax\n va_s0) inout_b len (va_get_mem_layout va_s0) Secret /\\ len == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 inout_b /\\ va_get_xmm 9 va_s0 == Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\ aes_reqs alg key round_keys\n keys_b (va_get_reg64 rR8 va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\\n pclmulqdq_enabled /\\ Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0\n va_s0) hkeys_b) (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret /\\ va_get_xmm 8 va_s0 == Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.AES.GHash.ghash_incremental0 h_LE old_hash completed_quads) /\\ FStar.Seq.Base.length\n #quad32 completed_quads == total_bytes `op_Division` 16 /\\ total_bytes < 16 `op_Multiply`\n FStar.Seq.Base.length #quad32 completed_quads + 16 /\\ va_get_reg64 rR10 va_s0 == total_bytes\n `op_Modulus` 16 /\\ total_bytes `op_Modulus` 16 =!= 0 /\\ (0 < total_bytes /\\ total_bytes < 16\n `op_Multiply` Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes) /\\ 16 `op_Multiply`\n (Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes - 1) < total_bytes) /\\ (forall\n (va_x_mem:vale_heap) (va_x_rcx:nat64) (va_x_r11:nat64) (va_x_xmm0:quad32) (va_x_xmm1:quad32)\n (va_x_xmm2:quad32) (va_x_xmm3:quad32) (va_x_xmm4:quad32) (va_x_xmm5:quad32) (va_x_xmm6:quad32)\n (va_x_xmm7:quad32) (va_x_xmm8:quad32) (va_x_xmm10:quad32) (va_x_heap5:vale_heap)\n (va_x_efl:Vale.X64.Flags.t) . let va_sM = va_upd_flags va_x_efl (va_upd_mem_heaplet 5\n va_x_heap5 (va_upd_xmm 10 va_x_xmm10 (va_upd_xmm 8 va_x_xmm8 (va_upd_xmm 7 va_x_xmm7\n (va_upd_xmm 6 va_x_xmm6 (va_upd_xmm 5 va_x_xmm5 (va_upd_xmm 4 va_x_xmm4 (va_upd_xmm 3 va_x_xmm3\n (va_upd_xmm 2 va_x_xmm2 (va_upd_xmm 1 va_x_xmm1 (va_upd_xmm 0 va_x_xmm0 (va_upd_reg64 rR11\n va_x_r11 (va_upd_reg64 rRcx va_x_rcx (va_upd_mem va_x_mem va_s0)))))))))))))) in va_get_ok\n va_sM /\\ (let (len:(va_int_range 1 1)) = 1 in Vale.X64.Decls.modifies_buffer128 inout_b\n (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5 va_sM) /\\ Vale.AES.GCTR.gctr_partial alg len\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b) key (va_get_xmm 11 va_s0) /\\ (let raw_quads =\n FStar.Seq.Base.append #quad32 completed_quads (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0)\n inout_b) in let input_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads) 0 total_bytes in let padded_bytes =\n Vale.AES.GCTR_s.pad_to_128_bits input_bytes in let input_quads =\n Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in l_and (FStar.Seq.Base.length\n #Vale.Def.Types_s.quad32 input_quads > 0) (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8\n va_sM) == Vale.AES.GHash.ghash_incremental h_LE old_hash input_quads))) ==> va_k va_sM (())))\n\n\nval va_wpProof_Gcm_extra_bytes : alg:algorithm -> inout_b:buffer128 -> key:(seq nat32) ->\n round_keys:(seq quad32) -> keys_b:buffer128 -> hkeys_b:buffer128 -> total_bytes:nat ->\n old_hash:quad32 -> completed_quads:(seq quad32) -> h_LE:quad32 -> va_s0:va_state ->\n va_k:(va_state -> unit -> Type0)\n -> Ghost (va_state & va_fuel & unit)\n (requires (va_t_require va_s0 /\\ va_wp_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b\n total_bytes old_hash completed_quads h_LE va_s0 va_k))\n (ensures (fun (va_sM, va_f0, va_g) -> va_t_ensure (va_code_Gcm_extra_bytes alg) ([va_Mod_flags;\n va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5;\n va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR11;\n va_Mod_reg64 rRcx; va_Mod_mem]) va_s0 va_k ((va_sM, va_f0, va_g))))\n[@\"opaque_to_smt\"]\nlet va_wpProof_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b total_bytes old_hash\n completed_quads h_LE va_s0 va_k =\n let (va_sM, va_f0) = va_lemma_Gcm_extra_bytes (va_code_Gcm_extra_bytes alg) va_s0 alg inout_b key\n round_keys keys_b hkeys_b total_bytes old_hash completed_quads h_LE in\n va_lemma_upd_update va_sM;\n assert (va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 5 va_sM (va_update_xmm 10\n va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5\n va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1\n va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR11 va_sM (va_update_reg64 rRcx va_sM\n (va_update_ok va_sM (va_update_mem va_sM va_s0)))))))))))))))));\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm\n 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1;\n va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_mem]) va_sM va_s0;\n let va_g = () in\n (va_sM, va_f0, va_g)\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_quick_Gcm_extra_bytes (alg:algorithm) (inout_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (total_bytes:nat) (old_hash:quad32)\n (completed_quads:(seq quad32)) (h_LE:quad32) : (va_quickCode unit (va_code_Gcm_extra_bytes alg)) =\n (va_QProc (va_code_Gcm_extra_bytes alg) ([va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10;\n va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm\n 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_mem])\n (va_wp_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b total_bytes old_hash\n completed_quads h_LE) (va_wpProof_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b\n total_bytes old_hash completed_quads h_LE))\n//--\n//-- Gcm_blocks128\n\nval va_code_Gcm_blocks128 : alg:algorithm -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_blocks128 alg =\n (va_Block (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRbx) (va_op_opr64_reg64 rRdi))\n (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR12) (va_op_opr64_reg64 rRdx)) (va_CCons\n (va_code_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRax)) (va_CCons\n (va_code_Ghash_buffer ()) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRdi)\n (va_op_opr64_reg64 rRbx)) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRdx)\n (va_op_opr64_reg64 rR12)) (va_CCons (va_code_Gctr_blocks128 alg) (va_CNil ())))))))))\n\nval va_codegen_success_Gcm_blocks128 : alg:algorithm -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_blocks128 alg =\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rRbx) (va_op_opr64_reg64 rRdi))\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR12) (va_op_opr64_reg64 rRdx))\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRax))\n (va_pbool_and (va_codegen_success_Ghash_buffer ()) (va_pbool_and (va_codegen_success_Mov64\n (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRbx)) (va_pbool_and (va_codegen_success_Mov64\n (va_op_dst_opr64_reg64 rRdx) (va_op_opr64_reg64 rR12)) (va_pbool_and\n (va_codegen_success_Gctr_blocks128 alg) (va_ttrue ()))))))))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_blocks128 (va_mods:va_mods_t) (alg:algorithm) (in_b:buffer128) (out_b:buffer128)\n (key:(seq nat32)) (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (h_LE:quad32)\n : (va_quickCode unit (va_code_Gcm_blocks128 alg)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 274 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRbx) (va_op_opr64_reg64 rRdi)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 275 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR12) (va_op_opr64_reg64 rRdx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 276 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRax)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 277 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Ghash_buffer hkeys_b in_b h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8\n va_old_s))) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 278 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRbx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 279 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRdx) (va_op_opr64_reg64 rR12)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 280 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gctr_blocks128 alg in_b out_b key round_keys keys_b) (va_QEmpty (()))))))))))\n\n\nval va_lemma_Gcm_blocks128 : va_b0:va_code -> va_s0:va_state -> alg:algorithm -> in_b:buffer128 ->\n out_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128 -> h_LE:quad32\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_blocks128 alg) va_s0 /\\ va_get_ok va_s0 /\\\n (sse_enabled /\\ Vale.X64.Decls.buffers_disjoint128 keys_b out_b /\\\n Vale.X64.Decls.buffers_disjoint128 hkeys_b out_b /\\ (Vale.X64.Decls.buffers_disjoint128 in_b\n out_b \\/ in_b == out_b) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0)\n (va_get_reg64 rRax va_s0) in_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1 va_s0) (va_get_reg64 rRdi va_s0) out_b\n (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\ va_get_reg64 rRax va_s0 + 16\n `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply`\n va_get_reg64 rRdx va_s0 < pow2_64 /\\ l_and (Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 in_b == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out_b)\n (Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b < pow2_32) /\\ va_get_reg64 rRdx\n va_s0 == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b /\\ va_get_xmm 9 va_s0 ==\n Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\\n va_get_reg64 rRdx va_s0 < pow2_32 /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rR8\n va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret)))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (Vale.X64.Decls.modifies_buffer128 out_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM) /\\ Vale.AES.GCTR.gctr_partial alg (va_get_reg64 rRdx va_s0) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b) key\n (va_get_xmm 11 va_s0) /\\ va_get_xmm 11 va_sM == Vale.AES.GCTR.inc32lite (va_get_xmm 11 va_s0)\n (va_get_reg64 rRdx va_s0) /\\ (va_get_reg64 rRdx va_s0 == 0 ==> l_and (va_get_xmm 8 va_sM ==\n va_get_xmm 8 va_s0) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b ==\n Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) out_b)) /\\ (va_get_reg64 rRdx va_s0 > 0 ==>\n l_and (va_get_reg64 rRdx va_s0 <= FStar.Seq.Base.length #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) ==> FStar.Seq.Base.length\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.slice #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) 0 (va_get_reg64 rRdx va_s0)) > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s0)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b)))) /\\ va_state_eq va_sM (va_update_flags va_sM\n (va_update_mem_heaplet 1 va_sM (va_update_xmm 10 va_sM (va_update_xmm 11 va_sM (va_update_xmm 8\n va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5 va_sM (va_update_xmm 4\n va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1 va_sM (va_update_xmm 0\n va_sM (va_update_reg64 rR12 va_sM (va_update_reg64 rRdx va_sM (va_update_reg64 rR10 va_sM\n (va_update_reg64 rR11 va_sM (va_update_reg64 rRdi va_sM (va_update_reg64 rRbx va_sM\n (va_update_ok va_sM (va_update_mem va_sM va_s0)))))))))))))))))))))))\n[@\"opaque_to_smt\"]\nlet va_lemma_Gcm_blocks128 va_b0 va_s0 alg in_b out_b key round_keys keys_b hkeys_b h_LE =\n let (va_mods:va_mods_t) = [va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11;\n va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm\n 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10;\n va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_ok; va_Mod_mem] in\n let va_qc = va_qcode_Gcm_blocks128 va_mods alg in_b out_b key round_keys keys_b hkeys_b h_LE in\n let (va_sM, va_fM, va_g) = va_wp_sound_code_norm (va_code_Gcm_blocks128 alg) va_qc va_s0 (fun\n va_s0 va_sM va_g -> let () = va_g in label va_range1\n \"***** POSTCONDITION NOT MET AT line 210 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\ (label va_range1\n \"***** POSTCONDITION NOT MET AT line 255 column 53 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 261 column 95 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.AES.GCTR.gctr_partial alg (va_get_reg64 rRdx va_s0) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b) key\n (va_get_xmm 11 va_s0)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 262 column 45 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_xmm 11 va_sM == Vale.AES.GCTR.inc32lite (va_get_xmm 11 va_s0) (va_get_reg64 rRdx\n va_s0)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 265 column 93 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRdx va_s0 == 0 ==> l_and (va_get_xmm 8 va_sM == va_get_xmm 8 va_s0)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b == Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) out_b)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 267 column 131 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRdx va_s0 > 0 ==> l_and (va_get_reg64 rRdx va_s0 <= FStar.Seq.Base.length\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) ==>\n FStar.Seq.Base.length #Vale.X64.Decls.quad32 (FStar.Seq.Base.slice #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) 0 (va_get_reg64 rRdx va_s0)) > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s0)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b))))) in\n assert_norm (va_qc.mods == va_mods);\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11; va_Mod_xmm\n 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2;\n va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10;\n va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_ok; va_Mod_mem]) va_sM va_s0;\n (va_sM, va_fM)\n\n[@ va_qattr]\nlet va_wp_Gcm_blocks128 (alg:algorithm) (in_b:buffer128) (out_b:buffer128) (key:(seq nat32))\n (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (h_LE:quad32) (va_s0:va_state)\n (va_k:(va_state -> unit -> Type0)) : Type0 =\n (va_get_ok va_s0 /\\ (sse_enabled /\\ Vale.X64.Decls.buffers_disjoint128 keys_b out_b /\\\n Vale.X64.Decls.buffers_disjoint128 hkeys_b out_b /\\ (Vale.X64.Decls.buffers_disjoint128 in_b\n out_b \\/ in_b == out_b) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0)\n (va_get_reg64 rRax va_s0) in_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1 va_s0) (va_get_reg64 rRdi va_s0) out_b\n (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\ va_get_reg64 rRax va_s0 + 16\n `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply`\n va_get_reg64 rRdx va_s0 < pow2_64 /\\ l_and (Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 in_b == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out_b)\n (Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b < pow2_32) /\\ va_get_reg64 rRdx\n va_s0 == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b /\\ va_get_xmm 9 va_s0 ==\n Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\\n va_get_reg64 rRdx va_s0 < pow2_32 /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rR8\n va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret) /\\ (forall (va_x_mem:vale_heap) (va_x_rbx:nat64) (va_x_rdi:nat64) (va_x_r11:nat64)\n (va_x_r10:nat64) (va_x_rdx:nat64) (va_x_r12:nat64) (va_x_xmm0:quad32) (va_x_xmm1:quad32)\n (va_x_xmm2:quad32) (va_x_xmm3:quad32) (va_x_xmm4:quad32) (va_x_xmm5:quad32) (va_x_xmm6:quad32)\n (va_x_xmm7:quad32) (va_x_xmm8:quad32) (va_x_xmm11:quad32) (va_x_xmm10:quad32)\n (va_x_heap1:vale_heap) (va_x_efl:Vale.X64.Flags.t) . let va_sM = va_upd_flags va_x_efl\n (va_upd_mem_heaplet 1 va_x_heap1 (va_upd_xmm 10 va_x_xmm10 (va_upd_xmm 11 va_x_xmm11\n (va_upd_xmm 8 va_x_xmm8 (va_upd_xmm 7 va_x_xmm7 (va_upd_xmm 6 va_x_xmm6 (va_upd_xmm 5 va_x_xmm5\n (va_upd_xmm 4 va_x_xmm4 (va_upd_xmm 3 va_x_xmm3 (va_upd_xmm 2 va_x_xmm2 (va_upd_xmm 1 va_x_xmm1\n (va_upd_xmm 0 va_x_xmm0 (va_upd_reg64 rR12 va_x_r12 (va_upd_reg64 rRdx va_x_rdx (va_upd_reg64\n rR10 va_x_r10 (va_upd_reg64 rR11 va_x_r11 (va_upd_reg64 rRdi va_x_rdi (va_upd_reg64 rRbx\n va_x_rbx (va_upd_mem va_x_mem va_s0))))))))))))))))))) in va_get_ok va_sM /\\\n (Vale.X64.Decls.modifies_buffer128 out_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM) /\\ Vale.AES.GCTR.gctr_partial alg (va_get_reg64 rRdx va_s0) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b) key\n (va_get_xmm 11 va_s0) /\\ va_get_xmm 11 va_sM == Vale.AES.GCTR.inc32lite (va_get_xmm 11 va_s0)\n (va_get_reg64 rRdx va_s0) /\\ (va_get_reg64 rRdx va_s0 == 0 ==> l_and (va_get_xmm 8 va_sM ==\n va_get_xmm 8 va_s0) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b ==\n Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) out_b)) /\\ (va_get_reg64 rRdx va_s0 > 0 ==>\n l_and (va_get_reg64 rRdx va_s0 <= FStar.Seq.Base.length #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) ==> FStar.Seq.Base.length\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.slice #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) 0 (va_get_reg64 rRdx va_s0)) > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s0)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b)))) ==> va_k va_sM (())))\n\n\nval va_wpProof_Gcm_blocks128 : alg:algorithm -> in_b:buffer128 -> out_b:buffer128 -> key:(seq\n nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 -> hkeys_b:buffer128 -> h_LE:quad32 ->\n va_s0:va_state -> va_k:(va_state -> unit -> Type0)\n -> Ghost (va_state & va_fuel & unit)\n (requires (va_t_require va_s0 /\\ va_wp_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b\n h_LE va_s0 va_k))\n (ensures (fun (va_sM, va_f0, va_g) -> va_t_ensure (va_code_Gcm_blocks128 alg) ([va_Mod_flags;\n va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6;\n va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0;\n va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10; va_Mod_reg64 rR11; va_Mod_reg64 rRdi;\n va_Mod_reg64 rRbx; va_Mod_mem]) va_s0 va_k ((va_sM, va_f0, va_g))))\n\nPartial implementation:\nlet va_wpProof_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b h_LE va_s0 va_k =\n \n"
},
{
"role": "assistant",
"content": "let va_sM, va_f0 =\n va_lemma_Gcm_blocks128 (va_code_Gcm_blocks128 alg) va_s0 alg in_b out_b key round_keys keys_b\n hkeys_b h_LE\nin\nva_lemma_upd_update va_sM;\nassert (va_state_eq va_sM\n (va_update_flags va_sM\n (va_update_mem_heaplet 1\n va_sM\n (va_update_xmm 10\n va_sM\n (va_update_xmm 11\n va_sM\n (va_update_xmm 8\n va_sM\n (va_update_xmm 7\n va_sM\n (va_update_xmm 6\n va_sM\n (va_update_xmm 5\n va_sM\n (va_update_xmm 4\n va_sM\n (va_update_xmm 3\n va_sM\n (va_update_xmm 2\n va_sM\n (va_update_xmm 1\n va_sM\n (va_update_xmm 0\n va_sM\n (va_update_reg64 rR12\n va_sM\n (va_update_reg64 rRdx\n va_sM\n (va_update_reg64 rR10\n va_sM\n (va_update_reg64 rR11\n va_sM\n (va_update_reg64 rRdi\n va_sM\n (va_update_reg64 rRbx\n va_sM\n (va_update_ok va_sM\n (va_update_mem\n va_sM\n va_s0)))))\n )))))))))))))))));\nva_lemma_norm_mods ([\n va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1;\n va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10; va_Mod_reg64 rR11;\n va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_mem\n ])\n va_sM\n va_s0;\nlet va_g = () in\n(va_sM, va_f0, va_g)"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: src1: Vale.X64.Machine_s.quad32 -> src2: Vale.X64.Machine_s.quad32\n -> FStar.Pervasives.Native.option Vale.X64.Machine_s.quad32\nEffect: Prims.Tot\n\nContext:\nmodule Vale.X64.Instructions_s\nopen FStar.Mul\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Two_s\nopen Vale.Def.Words.Four_s\nopen Vale.Def.Types_s\nopen Vale.X64.Machine_s\nopen Vale.X64.Instruction_s\nopen Vale.X64.CPU_Features_s\nopen Vale.X64.CryptoInstructions_s\nopen FStar.Seq.Base\n\nlet eval_Mov64 (src:nat64) : option nat64 = Some src\nval ins_Mov64 : instr_dep [out op64] [op64] PreserveFlags eval_Mov64\n\nlet eval_MovBe64 (src:nat64) : option nat64 = if movbe_enabled then Some (reverse_bytes_nat64 src) else None\nval ins_MovBe64 : instr_dep [out op64] [op64] PreserveFlags eval_MovBe64\n\nlet eval_Bswap64 (dst:nat64) : option nat64 = Some (reverse_bytes_nat64 dst)\nval ins_Bswap64 : instr_dep [inOut op64] [] PreserveFlags eval_Bswap64\n\nlet eval_Cmovc64 (dst src:nat64) (carry:bool) : option nat64 = Some (if carry then src else dst)\nval ins_Cmovc64 : instr_dep [inOut op64] [op64; opFlagsCf] PreserveFlags eval_Cmovc64\n\nlet eval_Add64 (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src in Some (sum >= pow2_64, sum % pow2_64)\nval ins_Add64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Add64\n\nlet eval_AddLea64 (src1 src2:nat64) : option nat64 = Some ((src1 + src2) % pow2_64)\nval ins_AddLea64 :instr_dep [out op64] [op64; op64] PreserveFlags eval_AddLea64\n\nlet eval_AddCarry64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_carry then 1 else 0) in\n Some (sum >= pow2_64, sum % pow2_64)\nval ins_AddCarry64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_AddCarry64\n\nlet eval_Adcx64_Adox64 (old_flag:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_flag then 1 else 0) in\n if adx_enabled then Some (sum >= pow2_64, sum % pow2_64) else None\nval ins_Adcx64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\nval ins_Adox64 : instr_dep [inOut opFlagsOf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\n\nlet eval_Sub64 (dst src:nat64) : option (bool & nat64) =\n let diff = dst - src in Some (diff < 0, diff % pow2_64)\nval ins_Sub64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Sub64\n\nlet eval_Sbb64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let diff = dst - (src + (if old_carry then 1 else 0)) in\n Some (diff < 0, diff % pow2_64) // We specify cf, but underspecify everything else (via HavocFlags)\nval ins_Sbb64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_Sbb64\n\nlet eval_Mul64 (rax src:nat64) : option (nat64 & nat64) =\n Some (FStar.UInt.mul_div #64 rax src, FStar.UInt.mul_mod #64 rax src)\nval ins_Mul64 : instr_dep [out (one64Reg rRdx); inOut (one64Reg rRax)] [op64] HavocFlags eval_Mul64\n\nlet eval_Mulx64 (rdx src:nat64) : option (nat64 & nat64) =\n let hi = FStar.UInt.mul_div #64 rdx src in\n let lo = FStar.UInt.mul_mod #64 rdx src in\n if bmi2_enabled then Some (hi, lo) else None\nval ins_Mulx64 : instr_dep [out op64; out op64] [one64Reg rRdx; op64] PreserveFlags eval_Mulx64\n\nlet eval_IMul64 (dst src:nat64) : option nat64 =\n Some (FStar.UInt.mul_mod #64 dst src)\nval ins_IMul64 : instr_dep [inOut op64] [op64] HavocFlags eval_IMul64\n\nlet eval_And64 (dst src:nat64) : option nat64 = Some (iand dst src)\nval ins_And64 : instr_dep [inOut op64] [op64] HavocFlags eval_And64\n\nlet eval_Xor64 (dst src:nat64) : option (nat64 & (bool & bool)) =\n Some (Vale.Def.Types_s.ixor dst src, (false, false))\nval ins_Xor64 : instr_dep [inOut op64; out opFlagsCf; out opFlagsOf] [op64] HavocFlags eval_Xor64\n\nlet eval_Shr64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishr dst amt) else None\nval ins_Shr64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shr64\n\nlet eval_Shl64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishl dst amt) else None\nval ins_Shl64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shl64\n\nlet eval_Cpuid (rax rcx:nat64) : option (nat64 & (nat64 & (nat64 & nat64))) =\n Some (cpuid rRax rax rcx, (cpuid rRbx rax rcx, (cpuid rRcx rax rcx, cpuid rRdx rax rcx)))\nval ins_Cpuid :\n instr_dep\n [inOut (one64Reg rRax); out (one64Reg rRbx); inOut (one64Reg rRcx); out (one64Reg rRdx)]\n [] PreserveFlags eval_Cpuid\n\n// The XGETBV instruction requires that OSXSAVE (in CPUID) is enabled.\n// We underspecify XGETBV here to only support fetching XCR0, which\n// is supported on any processor supporting the XGETBV instruction\nlet eval_Xgetbv (rcx:nat64) : option (nat64 & nat64) =\n if osxsave_enabled && rcx = 0 then Some (xgetbv rRax rcx, xgetbv rRdx rcx) else None\nval ins_Xgetbv :\n instr_dep\n [out (one64Reg rRax); out (one64Reg rRdx)]\n [one64Reg rRcx] PreserveFlags eval_Xgetbv\n\nlet check_avx (#a:Type0) (x:option a) : option a =\n if avx_enabled then x else None\n\nlet check_sse2 (#a:Type0) (x:option a) : option a =\n if sse2_enabled then x else None\n\nlet check_ssse3 (#a:Type0) (x:option a) : option a =\n if ssse3_enabled then x else None\n\nlet check_sse4_1 (#a:Type0) (x:option a) : option a =\n if sse4_1_enabled then x else None\n\nlet eval_Movdqu (src:quad32) : option quad32 = check_sse2 (Some src)\nval ins_Movdqu : instr_dep [out opXmm] [opXmm] PreserveFlags eval_Movdqu\n\nlet eval_Pxor (dst src:quad32) : option quad32 = check_sse2 (Some (quad32_xor dst src))\nval ins_Pxor : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pxor\n\nlet eval_VPxor (src1 src2:quad32) : option quad32 = check_avx (Some (quad32_xor src1 src2))\nval ins_VPxor : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPxor\n\nlet eval_Pand (dst src:quad32) : option quad32 = check_sse2 (Some (four_map2 (fun di si -> iand di si) dst src))\nval ins_Pand : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pand\n\nlet eval_Paddd_raw (src1 src2:quad32) : option quad32 =\n Some (Mkfour\n ((src1.lo0 + src2.lo0) % pow2_32)\n ((src1.lo1 + src2.lo1) % pow2_32)\n ((src1.hi2 + src2.hi2) % pow2_32)\n ((src1.hi3 + src2.hi3) % pow2_32))\n\nlet eval_Paddd (src1 src2:quad32) : option quad32 = check_sse2 (eval_Paddd_raw src1 src2)\nval ins_Paddd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Paddd\n\nlet eval_VPaddd (src1 src2:quad32) : option quad32 = check_avx (eval_Paddd_raw src1 src2)\nval ins_VPaddd : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPaddd\n\nlet eval_Pslld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishl i amt) dst) else None)\nval ins_Pslld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Pslld amt)\n\nlet eval_Psrld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishr i amt) dst) else None)\nval ins_Psrld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrld amt)\n\nlet eval_Psrldq (amt:int) (dst:quad32) : option quad32 = check_sse2 (\n if 0 <= amt && amt < 16 then\n let src_bytes = le_quad32_to_bytes dst in\n let zero_pad = Seq.create amt 0 in\n let remaining_bytes = slice src_bytes amt (length src_bytes) in\n Some (le_bytes_to_quad32 (append zero_pad remaining_bytes))\n else None)\nval ins_Psrldq (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrldq amt)\n\nlet eval_Palignr_raw (amount:nat8) (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if amount = 4 then\n Some (Mkfour src2.lo1 src2.hi2 src2.hi3 src1.lo0)\n else if amount = 8 then\n Some (Mkfour src2.hi2 src2.hi3 src1.lo0 src1.lo1)\n else None\n\nlet eval_Palignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_ssse3 (eval_Palignr_raw amount src1 src2)\nval ins_Palignr (amount:nat8) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Palignr amount)\n\nlet eval_VPalignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Palignr_raw amount src1 src2)\nval ins_VPalignr (amount:nat8) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPalignr amount)\n\nlet eval_Shufpd_raw (permutation:int) (src1 src2:quad32) : option quad32 =\n if 0 <= permutation && permutation < 4 then\n Some (Mkfour\n (if permutation % 2 = 0 then src1.lo0 else src1.hi2)\n (if permutation % 2 = 0 then src1.lo1 else src1.hi3)\n (if (permutation / 2) % 2 = 0 then src2.lo0 else src2.hi2)\n (if (permutation / 2) % 2 = 0 then src2.lo1 else src2.hi3))\n else None\n\nlet eval_Shufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_sse2 (eval_Shufpd_raw permutation src1 src2)\nval ins_Shufpd (permutation:int) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Shufpd permutation)\n\nlet eval_VShufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Shufpd_raw permutation src1 src2)\nval ins_VShufpd (permutation:int) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VShufpd permutation)\n\nlet is_full_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x04050607 &&\n q.hi3 = 0x00010203\n\nlet is_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x00010203 &&\n q.lo1 = 0x04050607 &&\n q.hi2 = 0x08090A0B &&\n q.hi3 = 0x0C0D0E0F\n\nlet is_high_dup_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet is_lower_upper_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x04050607 &&\n q.lo1 = 0x00010203 &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet eval_Pshufb_raw (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if is_full_byte_reversal_mask src2 then\n Some (reverse_bytes_quad32 src1)\n else if is_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3))\n else if is_high_dup_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else if is_lower_upper_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else None\n\nlet eval_Pshufb (src1 src2:quad32) : option quad32 = check_ssse3 (eval_Pshufb_raw src1 src2)\nval ins_Pshufb : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pshufb\n\nlet eval_VPshufb (src1 src2:quad32) : option quad32 = check_avx (eval_Pshufb_raw src1 src2)\nval ins_VPshufb : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPshufb\n\nlet eval_Pshufd (permutation:nat8) (src:quad32) : option quad32 = check_sse2 (\n let bits:bits_of_byte = byte_to_twobits permutation in\n Some (Mkfour\n (select_word src bits.lo0)\n (select_word src bits.lo1)\n (select_word src bits.hi2)\n (select_word src bits.hi3)))\nval ins_Pshufd (permutation:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_Pshufd permutation)\n\nlet eval_Pcmpeqd (dst src:quad32) : option quad32 = check_sse2 (\n let eq_result (b:bool):nat32 = if b then 0xFFFFFFFF else 0 in\n Some (Mkfour\n (eq_result (src.lo0 = dst.lo0))\n (eq_result (src.lo1 = dst.lo1))\n (eq_result (src.hi2 = dst.hi2))\n (eq_result (src.hi3 = dst.hi3))))\nval ins_Pcmpeqd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pcmpeqd\n\nlet eval_Pextrq (index:nat8) (src:quad32) : option nat64 = check_sse4_1 (\n let src_two = four_to_two_two src in\n Some (two_to_nat 32 (two_select src_two (index % 2))))\nval ins_Pextrq (index:nat8) : instr_dep [out op64] [opXmm] PreserveFlags (eval_Pextrq index)\n\nlet eval_Pinsrd (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat32 dst (src % pow2_32) (index % 4)))\nval ins_Pinsrd (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrd index)\n\nlet eval_Pinsrq (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat64_def dst src (index % 2)))\nval ins_Pinsrq (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrq index)\n\nlet eval_Pslldq_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the two very special cases we need\n if count = 4 then Some (Mkfour 0 src.lo0 src.lo1 src.hi2)\n else if count = 8 then Some (Mkfour 0 0 src.lo0 src.lo1)\n else None\nlet eval_VPslldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Pslldq_raw count src)\nval ins_VPslldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPslldq count)\n\nlet eval_Psrldq_8_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the one very special case we need\n if count = 8 then Some (Mkfour src.hi2 src.hi3 0 0)\n else None\nlet eval_VPsrldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Psrldq_8_raw count src)\nval ins_VPsrldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPsrldq count)\n\nlet eval_Pclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n let Mkfour a0 a1 a2 a3 = src1 in\n let Mkfour b0 b1 b2 b3 = src2 in\n let f x0 x1 y0 y1 =\n let x = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo x0 x1) in\n let y = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo y0 y1) in\n Vale.Math.Poly2.Bits_s.to_quad32 (Vale.Math.Poly2_s.mul x y)\n in\n if pclmulqdq_enabled then\n match imm with\n | 0 -> Some (f a0 a1 b0 b1)\n | 1 -> Some (f a2 a3 b0 b1)\n | 16 -> Some (f a0 a1 b2 b3)\n | 17 -> Some (f a2 a3 b2 b3)\n | _ -> None\n else None\nval ins_Pclmulqdq (imm:int) : instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Pclmulqdq imm)\n\nlet eval_VPclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Pclmulqdq imm src1 src2)\nval ins_VPclmulqdq (imm:int) : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPclmulqdq imm)\n\nlet eval_AESNI_enc (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.mix_columns_LE (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1))) src2)\n else None\nval ins_AESNI_enc : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc\n\nPartial implementation:\nlet eval_VAESNI_enc (src1 src2: quad32) : option quad32 =\n \n"
},
{
"role": "assistant",
"content": "check_avx (eval_AESNI_enc src1 src2)"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: count: Vale.Def.Types_s.nat8 -> src: Vale.X64.Machine_s.quad32\n -> FStar.Pervasives.Native.option Vale.X64.Machine_s.quad32\nEffect: Prims.Tot\n\nContext:\nmodule Vale.X64.Instructions_s\nopen FStar.Mul\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Two_s\nopen Vale.Def.Words.Four_s\nopen Vale.Def.Types_s\nopen Vale.X64.Machine_s\nopen Vale.X64.Instruction_s\nopen Vale.X64.CPU_Features_s\nopen Vale.X64.CryptoInstructions_s\nopen FStar.Seq.Base\n\nlet eval_Mov64 (src:nat64) : option nat64 = Some src\nval ins_Mov64 : instr_dep [out op64] [op64] PreserveFlags eval_Mov64\n\nlet eval_MovBe64 (src:nat64) : option nat64 = if movbe_enabled then Some (reverse_bytes_nat64 src) else None\nval ins_MovBe64 : instr_dep [out op64] [op64] PreserveFlags eval_MovBe64\n\nlet eval_Bswap64 (dst:nat64) : option nat64 = Some (reverse_bytes_nat64 dst)\nval ins_Bswap64 : instr_dep [inOut op64] [] PreserveFlags eval_Bswap64\n\nlet eval_Cmovc64 (dst src:nat64) (carry:bool) : option nat64 = Some (if carry then src else dst)\nval ins_Cmovc64 : instr_dep [inOut op64] [op64; opFlagsCf] PreserveFlags eval_Cmovc64\n\nlet eval_Add64 (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src in Some (sum >= pow2_64, sum % pow2_64)\nval ins_Add64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Add64\n\nlet eval_AddLea64 (src1 src2:nat64) : option nat64 = Some ((src1 + src2) % pow2_64)\nval ins_AddLea64 :instr_dep [out op64] [op64; op64] PreserveFlags eval_AddLea64\n\nlet eval_AddCarry64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_carry then 1 else 0) in\n Some (sum >= pow2_64, sum % pow2_64)\nval ins_AddCarry64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_AddCarry64\n\nlet eval_Adcx64_Adox64 (old_flag:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_flag then 1 else 0) in\n if adx_enabled then Some (sum >= pow2_64, sum % pow2_64) else None\nval ins_Adcx64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\nval ins_Adox64 : instr_dep [inOut opFlagsOf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\n\nlet eval_Sub64 (dst src:nat64) : option (bool & nat64) =\n let diff = dst - src in Some (diff < 0, diff % pow2_64)\nval ins_Sub64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Sub64\n\nlet eval_Sbb64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let diff = dst - (src + (if old_carry then 1 else 0)) in\n Some (diff < 0, diff % pow2_64) // We specify cf, but underspecify everything else (via HavocFlags)\nval ins_Sbb64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_Sbb64\n\nlet eval_Mul64 (rax src:nat64) : option (nat64 & nat64) =\n Some (FStar.UInt.mul_div #64 rax src, FStar.UInt.mul_mod #64 rax src)\nval ins_Mul64 : instr_dep [out (one64Reg rRdx); inOut (one64Reg rRax)] [op64] HavocFlags eval_Mul64\n\nlet eval_Mulx64 (rdx src:nat64) : option (nat64 & nat64) =\n let hi = FStar.UInt.mul_div #64 rdx src in\n let lo = FStar.UInt.mul_mod #64 rdx src in\n if bmi2_enabled then Some (hi, lo) else None\nval ins_Mulx64 : instr_dep [out op64; out op64] [one64Reg rRdx; op64] PreserveFlags eval_Mulx64\n\nlet eval_IMul64 (dst src:nat64) : option nat64 =\n Some (FStar.UInt.mul_mod #64 dst src)\nval ins_IMul64 : instr_dep [inOut op64] [op64] HavocFlags eval_IMul64\n\nlet eval_And64 (dst src:nat64) : option nat64 = Some (iand dst src)\nval ins_And64 : instr_dep [inOut op64] [op64] HavocFlags eval_And64\n\nlet eval_Xor64 (dst src:nat64) : option (nat64 & (bool & bool)) =\n Some (Vale.Def.Types_s.ixor dst src, (false, false))\nval ins_Xor64 : instr_dep [inOut op64; out opFlagsCf; out opFlagsOf] [op64] HavocFlags eval_Xor64\n\nlet eval_Shr64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishr dst amt) else None\nval ins_Shr64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shr64\n\nlet eval_Shl64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishl dst amt) else None\nval ins_Shl64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shl64\n\nlet eval_Cpuid (rax rcx:nat64) : option (nat64 & (nat64 & (nat64 & nat64))) =\n Some (cpuid rRax rax rcx, (cpuid rRbx rax rcx, (cpuid rRcx rax rcx, cpuid rRdx rax rcx)))\nval ins_Cpuid :\n instr_dep\n [inOut (one64Reg rRax); out (one64Reg rRbx); inOut (one64Reg rRcx); out (one64Reg rRdx)]\n [] PreserveFlags eval_Cpuid\n\n// The XGETBV instruction requires that OSXSAVE (in CPUID) is enabled.\n// We underspecify XGETBV here to only support fetching XCR0, which\n// is supported on any processor supporting the XGETBV instruction\nlet eval_Xgetbv (rcx:nat64) : option (nat64 & nat64) =\n if osxsave_enabled && rcx = 0 then Some (xgetbv rRax rcx, xgetbv rRdx rcx) else None\nval ins_Xgetbv :\n instr_dep\n [out (one64Reg rRax); out (one64Reg rRdx)]\n [one64Reg rRcx] PreserveFlags eval_Xgetbv\n\nlet check_avx (#a:Type0) (x:option a) : option a =\n if avx_enabled then x else None\n\nlet check_sse2 (#a:Type0) (x:option a) : option a =\n if sse2_enabled then x else None\n\nlet check_ssse3 (#a:Type0) (x:option a) : option a =\n if ssse3_enabled then x else None\n\nlet check_sse4_1 (#a:Type0) (x:option a) : option a =\n if sse4_1_enabled then x else None\n\nlet eval_Movdqu (src:quad32) : option quad32 = check_sse2 (Some src)\nval ins_Movdqu : instr_dep [out opXmm] [opXmm] PreserveFlags eval_Movdqu\n\nlet eval_Pxor (dst src:quad32) : option quad32 = check_sse2 (Some (quad32_xor dst src))\nval ins_Pxor : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pxor\n\nlet eval_VPxor (src1 src2:quad32) : option quad32 = check_avx (Some (quad32_xor src1 src2))\nval ins_VPxor : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPxor\n\nlet eval_Pand (dst src:quad32) : option quad32 = check_sse2 (Some (four_map2 (fun di si -> iand di si) dst src))\nval ins_Pand : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pand\n\nlet eval_Paddd_raw (src1 src2:quad32) : option quad32 =\n Some (Mkfour\n ((src1.lo0 + src2.lo0) % pow2_32)\n ((src1.lo1 + src2.lo1) % pow2_32)\n ((src1.hi2 + src2.hi2) % pow2_32)\n ((src1.hi3 + src2.hi3) % pow2_32))\n\nlet eval_Paddd (src1 src2:quad32) : option quad32 = check_sse2 (eval_Paddd_raw src1 src2)\nval ins_Paddd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Paddd\n\nlet eval_VPaddd (src1 src2:quad32) : option quad32 = check_avx (eval_Paddd_raw src1 src2)\nval ins_VPaddd : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPaddd\n\nlet eval_Pslld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishl i amt) dst) else None)\nval ins_Pslld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Pslld amt)\n\nlet eval_Psrld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishr i amt) dst) else None)\nval ins_Psrld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrld amt)\n\nlet eval_Psrldq (amt:int) (dst:quad32) : option quad32 = check_sse2 (\n if 0 <= amt && amt < 16 then\n let src_bytes = le_quad32_to_bytes dst in\n let zero_pad = Seq.create amt 0 in\n let remaining_bytes = slice src_bytes amt (length src_bytes) in\n Some (le_bytes_to_quad32 (append zero_pad remaining_bytes))\n else None)\nval ins_Psrldq (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrldq amt)\n\nlet eval_Palignr_raw (amount:nat8) (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if amount = 4 then\n Some (Mkfour src2.lo1 src2.hi2 src2.hi3 src1.lo0)\n else if amount = 8 then\n Some (Mkfour src2.hi2 src2.hi3 src1.lo0 src1.lo1)\n else None\n\nlet eval_Palignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_ssse3 (eval_Palignr_raw amount src1 src2)\nval ins_Palignr (amount:nat8) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Palignr amount)\n\nlet eval_VPalignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Palignr_raw amount src1 src2)\nval ins_VPalignr (amount:nat8) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPalignr amount)\n\nlet eval_Shufpd_raw (permutation:int) (src1 src2:quad32) : option quad32 =\n if 0 <= permutation && permutation < 4 then\n Some (Mkfour\n (if permutation % 2 = 0 then src1.lo0 else src1.hi2)\n (if permutation % 2 = 0 then src1.lo1 else src1.hi3)\n (if (permutation / 2) % 2 = 0 then src2.lo0 else src2.hi2)\n (if (permutation / 2) % 2 = 0 then src2.lo1 else src2.hi3))\n else None\n\nlet eval_Shufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_sse2 (eval_Shufpd_raw permutation src1 src2)\nval ins_Shufpd (permutation:int) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Shufpd permutation)\n\nlet eval_VShufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Shufpd_raw permutation src1 src2)\nval ins_VShufpd (permutation:int) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VShufpd permutation)\n\nlet is_full_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x04050607 &&\n q.hi3 = 0x00010203\n\nlet is_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x00010203 &&\n q.lo1 = 0x04050607 &&\n q.hi2 = 0x08090A0B &&\n q.hi3 = 0x0C0D0E0F\n\nlet is_high_dup_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet is_lower_upper_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x04050607 &&\n q.lo1 = 0x00010203 &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet eval_Pshufb_raw (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if is_full_byte_reversal_mask src2 then\n Some (reverse_bytes_quad32 src1)\n else if is_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3))\n else if is_high_dup_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else if is_lower_upper_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else None\n\nlet eval_Pshufb (src1 src2:quad32) : option quad32 = check_ssse3 (eval_Pshufb_raw src1 src2)\nval ins_Pshufb : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pshufb\n\nlet eval_VPshufb (src1 src2:quad32) : option quad32 = check_avx (eval_Pshufb_raw src1 src2)\nval ins_VPshufb : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPshufb\n\nlet eval_Pshufd (permutation:nat8) (src:quad32) : option quad32 = check_sse2 (\n let bits:bits_of_byte = byte_to_twobits permutation in\n Some (Mkfour\n (select_word src bits.lo0)\n (select_word src bits.lo1)\n (select_word src bits.hi2)\n (select_word src bits.hi3)))\nval ins_Pshufd (permutation:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_Pshufd permutation)\n\nlet eval_Pcmpeqd (dst src:quad32) : option quad32 = check_sse2 (\n let eq_result (b:bool):nat32 = if b then 0xFFFFFFFF else 0 in\n Some (Mkfour\n (eq_result (src.lo0 = dst.lo0))\n (eq_result (src.lo1 = dst.lo1))\n (eq_result (src.hi2 = dst.hi2))\n (eq_result (src.hi3 = dst.hi3))))\nval ins_Pcmpeqd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pcmpeqd\n\nlet eval_Pextrq (index:nat8) (src:quad32) : option nat64 = check_sse4_1 (\n let src_two = four_to_two_two src in\n Some (two_to_nat 32 (two_select src_two (index % 2))))\nval ins_Pextrq (index:nat8) : instr_dep [out op64] [opXmm] PreserveFlags (eval_Pextrq index)\n\nlet eval_Pinsrd (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat32 dst (src % pow2_32) (index % 4)))\nval ins_Pinsrd (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrd index)\n\nlet eval_Pinsrq (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat64_def dst src (index % 2)))\nval ins_Pinsrq (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrq index)\n\nlet eval_Pslldq_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the two very special cases we need\n if count = 4 then Some (Mkfour 0 src.lo0 src.lo1 src.hi2)\n else if count = 8 then Some (Mkfour 0 0 src.lo0 src.lo1)\n else None\nlet eval_VPslldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Pslldq_raw count src)\nval ins_VPslldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPslldq count)\n\nPartial implementation:\nlet eval_Psrldq_8_raw (count: nat8) (src: quad32) : option quad32 =\n \n"
},
{
"role": "assistant",
"content": "if count = 8 then Some (Mkfour src.hi2 src.hi3 0 0) else None"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: FStar.Pervasives.Native.option Prims.unit\nEffect: Prims.Tot\n\nContext:\nmodule Vale.X64.Instructions_s\nopen FStar.Mul\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Two_s\nopen Vale.Def.Words.Four_s\nopen Vale.Def.Types_s\nopen Vale.X64.Machine_s\nopen Vale.X64.Instruction_s\nopen Vale.X64.CPU_Features_s\nopen Vale.X64.CryptoInstructions_s\nopen FStar.Seq.Base\n\nlet eval_Mov64 (src:nat64) : option nat64 = Some src\nval ins_Mov64 : instr_dep [out op64] [op64] PreserveFlags eval_Mov64\n\nlet eval_MovBe64 (src:nat64) : option nat64 = if movbe_enabled then Some (reverse_bytes_nat64 src) else None\nval ins_MovBe64 : instr_dep [out op64] [op64] PreserveFlags eval_MovBe64\n\nlet eval_Bswap64 (dst:nat64) : option nat64 = Some (reverse_bytes_nat64 dst)\nval ins_Bswap64 : instr_dep [inOut op64] [] PreserveFlags eval_Bswap64\n\nlet eval_Cmovc64 (dst src:nat64) (carry:bool) : option nat64 = Some (if carry then src else dst)\nval ins_Cmovc64 : instr_dep [inOut op64] [op64; opFlagsCf] PreserveFlags eval_Cmovc64\n\nlet eval_Add64 (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src in Some (sum >= pow2_64, sum % pow2_64)\nval ins_Add64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Add64\n\nlet eval_AddLea64 (src1 src2:nat64) : option nat64 = Some ((src1 + src2) % pow2_64)\nval ins_AddLea64 :instr_dep [out op64] [op64; op64] PreserveFlags eval_AddLea64\n\nlet eval_AddCarry64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_carry then 1 else 0) in\n Some (sum >= pow2_64, sum % pow2_64)\nval ins_AddCarry64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_AddCarry64\n\nlet eval_Adcx64_Adox64 (old_flag:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_flag then 1 else 0) in\n if adx_enabled then Some (sum >= pow2_64, sum % pow2_64) else None\nval ins_Adcx64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\nval ins_Adox64 : instr_dep [inOut opFlagsOf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\n\nlet eval_Sub64 (dst src:nat64) : option (bool & nat64) =\n let diff = dst - src in Some (diff < 0, diff % pow2_64)\nval ins_Sub64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Sub64\n\nlet eval_Sbb64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let diff = dst - (src + (if old_carry then 1 else 0)) in\n Some (diff < 0, diff % pow2_64) // We specify cf, but underspecify everything else (via HavocFlags)\nval ins_Sbb64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_Sbb64\n\nlet eval_Mul64 (rax src:nat64) : option (nat64 & nat64) =\n Some (FStar.UInt.mul_div #64 rax src, FStar.UInt.mul_mod #64 rax src)\nval ins_Mul64 : instr_dep [out (one64Reg rRdx); inOut (one64Reg rRax)] [op64] HavocFlags eval_Mul64\n\nlet eval_Mulx64 (rdx src:nat64) : option (nat64 & nat64) =\n let hi = FStar.UInt.mul_div #64 rdx src in\n let lo = FStar.UInt.mul_mod #64 rdx src in\n if bmi2_enabled then Some (hi, lo) else None\nval ins_Mulx64 : instr_dep [out op64; out op64] [one64Reg rRdx; op64] PreserveFlags eval_Mulx64\n\nlet eval_IMul64 (dst src:nat64) : option nat64 =\n Some (FStar.UInt.mul_mod #64 dst src)\nval ins_IMul64 : instr_dep [inOut op64] [op64] HavocFlags eval_IMul64\n\nlet eval_And64 (dst src:nat64) : option nat64 = Some (iand dst src)\nval ins_And64 : instr_dep [inOut op64] [op64] HavocFlags eval_And64\n\nlet eval_Xor64 (dst src:nat64) : option (nat64 & (bool & bool)) =\n Some (Vale.Def.Types_s.ixor dst src, (false, false))\nval ins_Xor64 : instr_dep [inOut op64; out opFlagsCf; out opFlagsOf] [op64] HavocFlags eval_Xor64\n\nlet eval_Shr64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishr dst amt) else None\nval ins_Shr64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shr64\n\nlet eval_Shl64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishl dst amt) else None\nval ins_Shl64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shl64\n\nlet eval_Cpuid (rax rcx:nat64) : option (nat64 & (nat64 & (nat64 & nat64))) =\n Some (cpuid rRax rax rcx, (cpuid rRbx rax rcx, (cpuid rRcx rax rcx, cpuid rRdx rax rcx)))\nval ins_Cpuid :\n instr_dep\n [inOut (one64Reg rRax); out (one64Reg rRbx); inOut (one64Reg rRcx); out (one64Reg rRdx)]\n [] PreserveFlags eval_Cpuid\n\n// The XGETBV instruction requires that OSXSAVE (in CPUID) is enabled.\n// We underspecify XGETBV here to only support fetching XCR0, which\n// is supported on any processor supporting the XGETBV instruction\nlet eval_Xgetbv (rcx:nat64) : option (nat64 & nat64) =\n if osxsave_enabled && rcx = 0 then Some (xgetbv rRax rcx, xgetbv rRdx rcx) else None\nval ins_Xgetbv :\n instr_dep\n [out (one64Reg rRax); out (one64Reg rRdx)]\n [one64Reg rRcx] PreserveFlags eval_Xgetbv\n\nlet check_avx (#a:Type0) (x:option a) : option a =\n if avx_enabled then x else None\n\nlet check_sse2 (#a:Type0) (x:option a) : option a =\n if sse2_enabled then x else None\n\nlet check_ssse3 (#a:Type0) (x:option a) : option a =\n if ssse3_enabled then x else None\n\nlet check_sse4_1 (#a:Type0) (x:option a) : option a =\n if sse4_1_enabled then x else None\n\nlet eval_Movdqu (src:quad32) : option quad32 = check_sse2 (Some src)\nval ins_Movdqu : instr_dep [out opXmm] [opXmm] PreserveFlags eval_Movdqu\n\nlet eval_Pxor (dst src:quad32) : option quad32 = check_sse2 (Some (quad32_xor dst src))\nval ins_Pxor : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pxor\n\nlet eval_VPxor (src1 src2:quad32) : option quad32 = check_avx (Some (quad32_xor src1 src2))\nval ins_VPxor : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPxor\n\nlet eval_Pand (dst src:quad32) : option quad32 = check_sse2 (Some (four_map2 (fun di si -> iand di si) dst src))\nval ins_Pand : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pand\n\nlet eval_Paddd_raw (src1 src2:quad32) : option quad32 =\n Some (Mkfour\n ((src1.lo0 + src2.lo0) % pow2_32)\n ((src1.lo1 + src2.lo1) % pow2_32)\n ((src1.hi2 + src2.hi2) % pow2_32)\n ((src1.hi3 + src2.hi3) % pow2_32))\n\nlet eval_Paddd (src1 src2:quad32) : option quad32 = check_sse2 (eval_Paddd_raw src1 src2)\nval ins_Paddd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Paddd\n\nlet eval_VPaddd (src1 src2:quad32) : option quad32 = check_avx (eval_Paddd_raw src1 src2)\nval ins_VPaddd : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPaddd\n\nlet eval_Pslld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishl i amt) dst) else None)\nval ins_Pslld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Pslld amt)\n\nlet eval_Psrld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishr i amt) dst) else None)\nval ins_Psrld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrld amt)\n\nlet eval_Psrldq (amt:int) (dst:quad32) : option quad32 = check_sse2 (\n if 0 <= amt && amt < 16 then\n let src_bytes = le_quad32_to_bytes dst in\n let zero_pad = Seq.create amt 0 in\n let remaining_bytes = slice src_bytes amt (length src_bytes) in\n Some (le_bytes_to_quad32 (append zero_pad remaining_bytes))\n else None)\nval ins_Psrldq (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrldq amt)\n\nlet eval_Palignr_raw (amount:nat8) (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if amount = 4 then\n Some (Mkfour src2.lo1 src2.hi2 src2.hi3 src1.lo0)\n else if amount = 8 then\n Some (Mkfour src2.hi2 src2.hi3 src1.lo0 src1.lo1)\n else None\n\nlet eval_Palignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_ssse3 (eval_Palignr_raw amount src1 src2)\nval ins_Palignr (amount:nat8) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Palignr amount)\n\nlet eval_VPalignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Palignr_raw amount src1 src2)\nval ins_VPalignr (amount:nat8) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPalignr amount)\n\nlet eval_Shufpd_raw (permutation:int) (src1 src2:quad32) : option quad32 =\n if 0 <= permutation && permutation < 4 then\n Some (Mkfour\n (if permutation % 2 = 0 then src1.lo0 else src1.hi2)\n (if permutation % 2 = 0 then src1.lo1 else src1.hi3)\n (if (permutation / 2) % 2 = 0 then src2.lo0 else src2.hi2)\n (if (permutation / 2) % 2 = 0 then src2.lo1 else src2.hi3))\n else None\n\nlet eval_Shufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_sse2 (eval_Shufpd_raw permutation src1 src2)\nval ins_Shufpd (permutation:int) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Shufpd permutation)\n\nlet eval_VShufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Shufpd_raw permutation src1 src2)\nval ins_VShufpd (permutation:int) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VShufpd permutation)\n\nlet is_full_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x04050607 &&\n q.hi3 = 0x00010203\n\nlet is_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x00010203 &&\n q.lo1 = 0x04050607 &&\n q.hi2 = 0x08090A0B &&\n q.hi3 = 0x0C0D0E0F\n\nlet is_high_dup_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet is_lower_upper_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x04050607 &&\n q.lo1 = 0x00010203 &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet eval_Pshufb_raw (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if is_full_byte_reversal_mask src2 then\n Some (reverse_bytes_quad32 src1)\n else if is_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3))\n else if is_high_dup_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else if is_lower_upper_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else None\n\nlet eval_Pshufb (src1 src2:quad32) : option quad32 = check_ssse3 (eval_Pshufb_raw src1 src2)\nval ins_Pshufb : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pshufb\n\nlet eval_VPshufb (src1 src2:quad32) : option quad32 = check_avx (eval_Pshufb_raw src1 src2)\nval ins_VPshufb : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPshufb\n\nlet eval_Pshufd (permutation:nat8) (src:quad32) : option quad32 = check_sse2 (\n let bits:bits_of_byte = byte_to_twobits permutation in\n Some (Mkfour\n (select_word src bits.lo0)\n (select_word src bits.lo1)\n (select_word src bits.hi2)\n (select_word src bits.hi3)))\nval ins_Pshufd (permutation:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_Pshufd permutation)\n\nlet eval_Pcmpeqd (dst src:quad32) : option quad32 = check_sse2 (\n let eq_result (b:bool):nat32 = if b then 0xFFFFFFFF else 0 in\n Some (Mkfour\n (eq_result (src.lo0 = dst.lo0))\n (eq_result (src.lo1 = dst.lo1))\n (eq_result (src.hi2 = dst.hi2))\n (eq_result (src.hi3 = dst.hi3))))\nval ins_Pcmpeqd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pcmpeqd\n\nlet eval_Pextrq (index:nat8) (src:quad32) : option nat64 = check_sse4_1 (\n let src_two = four_to_two_two src in\n Some (two_to_nat 32 (two_select src_two (index % 2))))\nval ins_Pextrq (index:nat8) : instr_dep [out op64] [opXmm] PreserveFlags (eval_Pextrq index)\n\nlet eval_Pinsrd (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat32 dst (src % pow2_32) (index % 4)))\nval ins_Pinsrd (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrd index)\n\nlet eval_Pinsrq (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat64_def dst src (index % 2)))\nval ins_Pinsrq (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrq index)\n\nlet eval_Pslldq_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the two very special cases we need\n if count = 4 then Some (Mkfour 0 src.lo0 src.lo1 src.hi2)\n else if count = 8 then Some (Mkfour 0 0 src.lo0 src.lo1)\n else None\nlet eval_VPslldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Pslldq_raw count src)\nval ins_VPslldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPslldq count)\n\nlet eval_Psrldq_8_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the one very special case we need\n if count = 8 then Some (Mkfour src.hi2 src.hi3 0 0)\n else None\nlet eval_VPsrldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Psrldq_8_raw count src)\nval ins_VPsrldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPsrldq count)\n\nlet eval_Pclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n let Mkfour a0 a1 a2 a3 = src1 in\n let Mkfour b0 b1 b2 b3 = src2 in\n let f x0 x1 y0 y1 =\n let x = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo x0 x1) in\n let y = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo y0 y1) in\n Vale.Math.Poly2.Bits_s.to_quad32 (Vale.Math.Poly2_s.mul x y)\n in\n if pclmulqdq_enabled then\n match imm with\n | 0 -> Some (f a0 a1 b0 b1)\n | 1 -> Some (f a2 a3 b0 b1)\n | 16 -> Some (f a0 a1 b2 b3)\n | 17 -> Some (f a2 a3 b2 b3)\n | _ -> None\n else None\nval ins_Pclmulqdq (imm:int) : instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Pclmulqdq imm)\n\nlet eval_VPclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Pclmulqdq imm src1 src2)\nval ins_VPclmulqdq (imm:int) : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPclmulqdq imm)\n\nlet eval_AESNI_enc (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.mix_columns_LE (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1))) src2)\n else None\nval ins_AESNI_enc : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc\n\nlet eval_VAESNI_enc (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc src1 src2)\nval ins_VAESNI_enc : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc\n\nlet eval_AESNI_enc_last (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1)) src2)\n else None\nval ins_AESNI_enc_last : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc_last\n\nlet eval_VAESNI_enc_last (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc_last src1 src2)\nval ins_VAESNI_enc_last : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc_last\n\nlet eval_AESNI_dec (dst src:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.inv_mix_columns_LE (Vale.AES.AES_s.inv_sub_bytes (Vale.AES.AES_s.inv_shift_rows_LE dst))) src)\n else None\nval ins_AESNI_dec : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_dec\n\nlet eval_AESNI_dec_last (dst src:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.inv_sub_bytes (Vale.AES.AES_s.inv_shift_rows_LE dst)) src)\n else None\nval ins_AESNI_dec_last : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_dec_last\n\nlet eval_AESNI_imc (src:quad32) : option quad32 =\n if aesni_enabled then Some (Vale.AES.AES_s.inv_mix_columns_LE src) else None\nval ins_AESNI_imc : instr_dep [out opXmm] [opXmm] PreserveFlags eval_AESNI_imc\n\nlet eval_AESNI_keygen_assist (imm:nat8) (src:quad32) : option quad32 =\n if aesni_enabled then\n Some (Mkfour\n (Vale.AES.AES_s.sub_word src.lo1)\n (ixor (Vale.AES.AES_s.rot_word_LE (Vale.AES.AES_s.sub_word src.lo1)) imm)\n (Vale.AES.AES_s.sub_word src.hi3)\n (ixor (Vale.AES.AES_s.rot_word_LE (Vale.AES.AES_s.sub_word src.hi3)) imm))\n else None\nval ins_AESNI_keygen_assist (imm:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_AESNI_keygen_assist imm)\n\nlet eval_SHA256_rnds2 (src1 src2 wk:quad32) : option quad32 =\n if sha_enabled then Some (sha256_rnds2_spec src1 src2 wk) else None\nval ins_SHA256_rnds2 :\n instr_dep [inOut opXmm] [opXmm; oneXmm (OReg 0)] PreserveFlags eval_SHA256_rnds2\n\nlet eval_SHA256_msg1 (src1 src2:quad32) : option quad32 =\n if sha_enabled then Some (sha256_msg1_spec src1 src2) else None\nval ins_SHA256_msg1 : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_SHA256_msg1\n\nlet eval_SHA256_msg2 (src1 src2:quad32) : option quad32 =\n if sha_enabled then Some (sha256_msg2_spec src1 src2) else None\nval ins_SHA256_msg2 : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_SHA256_msg2\n\nlet eval_Ghost : option unit = Some ()\nval ins_Ghost : instr_dep [] [] PreserveFlags eval_Ghost\n\nPartial implementation:\nlet eval_Comment:option unit =\n \n"
},
{
"role": "assistant",
"content": "Some ()"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: imm: Prims.int -> src1: Vale.X64.Machine_s.quad32 -> src2: Vale.X64.Machine_s.quad32\n -> FStar.Pervasives.Native.option Vale.X64.Machine_s.quad32\nEffect: Prims.Tot\n\nContext:\nmodule Vale.X64.Instructions_s\nopen FStar.Mul\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Two_s\nopen Vale.Def.Words.Four_s\nopen Vale.Def.Types_s\nopen Vale.X64.Machine_s\nopen Vale.X64.Instruction_s\nopen Vale.X64.CPU_Features_s\nopen Vale.X64.CryptoInstructions_s\nopen FStar.Seq.Base\n\nlet eval_Mov64 (src:nat64) : option nat64 = Some src\nval ins_Mov64 : instr_dep [out op64] [op64] PreserveFlags eval_Mov64\n\nlet eval_MovBe64 (src:nat64) : option nat64 = if movbe_enabled then Some (reverse_bytes_nat64 src) else None\nval ins_MovBe64 : instr_dep [out op64] [op64] PreserveFlags eval_MovBe64\n\nlet eval_Bswap64 (dst:nat64) : option nat64 = Some (reverse_bytes_nat64 dst)\nval ins_Bswap64 : instr_dep [inOut op64] [] PreserveFlags eval_Bswap64\n\nlet eval_Cmovc64 (dst src:nat64) (carry:bool) : option nat64 = Some (if carry then src else dst)\nval ins_Cmovc64 : instr_dep [inOut op64] [op64; opFlagsCf] PreserveFlags eval_Cmovc64\n\nlet eval_Add64 (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src in Some (sum >= pow2_64, sum % pow2_64)\nval ins_Add64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Add64\n\nlet eval_AddLea64 (src1 src2:nat64) : option nat64 = Some ((src1 + src2) % pow2_64)\nval ins_AddLea64 :instr_dep [out op64] [op64; op64] PreserveFlags eval_AddLea64\n\nlet eval_AddCarry64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_carry then 1 else 0) in\n Some (sum >= pow2_64, sum % pow2_64)\nval ins_AddCarry64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_AddCarry64\n\nlet eval_Adcx64_Adox64 (old_flag:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_flag then 1 else 0) in\n if adx_enabled then Some (sum >= pow2_64, sum % pow2_64) else None\nval ins_Adcx64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\nval ins_Adox64 : instr_dep [inOut opFlagsOf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\n\nlet eval_Sub64 (dst src:nat64) : option (bool & nat64) =\n let diff = dst - src in Some (diff < 0, diff % pow2_64)\nval ins_Sub64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Sub64\n\nlet eval_Sbb64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let diff = dst - (src + (if old_carry then 1 else 0)) in\n Some (diff < 0, diff % pow2_64) // We specify cf, but underspecify everything else (via HavocFlags)\nval ins_Sbb64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_Sbb64\n\nlet eval_Mul64 (rax src:nat64) : option (nat64 & nat64) =\n Some (FStar.UInt.mul_div #64 rax src, FStar.UInt.mul_mod #64 rax src)\nval ins_Mul64 : instr_dep [out (one64Reg rRdx); inOut (one64Reg rRax)] [op64] HavocFlags eval_Mul64\n\nlet eval_Mulx64 (rdx src:nat64) : option (nat64 & nat64) =\n let hi = FStar.UInt.mul_div #64 rdx src in\n let lo = FStar.UInt.mul_mod #64 rdx src in\n if bmi2_enabled then Some (hi, lo) else None\nval ins_Mulx64 : instr_dep [out op64; out op64] [one64Reg rRdx; op64] PreserveFlags eval_Mulx64\n\nlet eval_IMul64 (dst src:nat64) : option nat64 =\n Some (FStar.UInt.mul_mod #64 dst src)\nval ins_IMul64 : instr_dep [inOut op64] [op64] HavocFlags eval_IMul64\n\nlet eval_And64 (dst src:nat64) : option nat64 = Some (iand dst src)\nval ins_And64 : instr_dep [inOut op64] [op64] HavocFlags eval_And64\n\nlet eval_Xor64 (dst src:nat64) : option (nat64 & (bool & bool)) =\n Some (Vale.Def.Types_s.ixor dst src, (false, false))\nval ins_Xor64 : instr_dep [inOut op64; out opFlagsCf; out opFlagsOf] [op64] HavocFlags eval_Xor64\n\nlet eval_Shr64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishr dst amt) else None\nval ins_Shr64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shr64\n\nlet eval_Shl64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishl dst amt) else None\nval ins_Shl64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shl64\n\nlet eval_Cpuid (rax rcx:nat64) : option (nat64 & (nat64 & (nat64 & nat64))) =\n Some (cpuid rRax rax rcx, (cpuid rRbx rax rcx, (cpuid rRcx rax rcx, cpuid rRdx rax rcx)))\nval ins_Cpuid :\n instr_dep\n [inOut (one64Reg rRax); out (one64Reg rRbx); inOut (one64Reg rRcx); out (one64Reg rRdx)]\n [] PreserveFlags eval_Cpuid\n\n// The XGETBV instruction requires that OSXSAVE (in CPUID) is enabled.\n// We underspecify XGETBV here to only support fetching XCR0, which\n// is supported on any processor supporting the XGETBV instruction\nlet eval_Xgetbv (rcx:nat64) : option (nat64 & nat64) =\n if osxsave_enabled && rcx = 0 then Some (xgetbv rRax rcx, xgetbv rRdx rcx) else None\nval ins_Xgetbv :\n instr_dep\n [out (one64Reg rRax); out (one64Reg rRdx)]\n [one64Reg rRcx] PreserveFlags eval_Xgetbv\n\nlet check_avx (#a:Type0) (x:option a) : option a =\n if avx_enabled then x else None\n\nlet check_sse2 (#a:Type0) (x:option a) : option a =\n if sse2_enabled then x else None\n\nlet check_ssse3 (#a:Type0) (x:option a) : option a =\n if ssse3_enabled then x else None\n\nlet check_sse4_1 (#a:Type0) (x:option a) : option a =\n if sse4_1_enabled then x else None\n\nlet eval_Movdqu (src:quad32) : option quad32 = check_sse2 (Some src)\nval ins_Movdqu : instr_dep [out opXmm] [opXmm] PreserveFlags eval_Movdqu\n\nlet eval_Pxor (dst src:quad32) : option quad32 = check_sse2 (Some (quad32_xor dst src))\nval ins_Pxor : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pxor\n\nlet eval_VPxor (src1 src2:quad32) : option quad32 = check_avx (Some (quad32_xor src1 src2))\nval ins_VPxor : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPxor\n\nlet eval_Pand (dst src:quad32) : option quad32 = check_sse2 (Some (four_map2 (fun di si -> iand di si) dst src))\nval ins_Pand : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pand\n\nlet eval_Paddd_raw (src1 src2:quad32) : option quad32 =\n Some (Mkfour\n ((src1.lo0 + src2.lo0) % pow2_32)\n ((src1.lo1 + src2.lo1) % pow2_32)\n ((src1.hi2 + src2.hi2) % pow2_32)\n ((src1.hi3 + src2.hi3) % pow2_32))\n\nlet eval_Paddd (src1 src2:quad32) : option quad32 = check_sse2 (eval_Paddd_raw src1 src2)\nval ins_Paddd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Paddd\n\nlet eval_VPaddd (src1 src2:quad32) : option quad32 = check_avx (eval_Paddd_raw src1 src2)\nval ins_VPaddd : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPaddd\n\nlet eval_Pslld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishl i amt) dst) else None)\nval ins_Pslld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Pslld amt)\n\nlet eval_Psrld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishr i amt) dst) else None)\nval ins_Psrld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrld amt)\n\nlet eval_Psrldq (amt:int) (dst:quad32) : option quad32 = check_sse2 (\n if 0 <= amt && amt < 16 then\n let src_bytes = le_quad32_to_bytes dst in\n let zero_pad = Seq.create amt 0 in\n let remaining_bytes = slice src_bytes amt (length src_bytes) in\n Some (le_bytes_to_quad32 (append zero_pad remaining_bytes))\n else None)\nval ins_Psrldq (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrldq amt)\n\nlet eval_Palignr_raw (amount:nat8) (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if amount = 4 then\n Some (Mkfour src2.lo1 src2.hi2 src2.hi3 src1.lo0)\n else if amount = 8 then\n Some (Mkfour src2.hi2 src2.hi3 src1.lo0 src1.lo1)\n else None\n\nlet eval_Palignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_ssse3 (eval_Palignr_raw amount src1 src2)\nval ins_Palignr (amount:nat8) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Palignr amount)\n\nlet eval_VPalignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Palignr_raw amount src1 src2)\nval ins_VPalignr (amount:nat8) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPalignr amount)\n\nlet eval_Shufpd_raw (permutation:int) (src1 src2:quad32) : option quad32 =\n if 0 <= permutation && permutation < 4 then\n Some (Mkfour\n (if permutation % 2 = 0 then src1.lo0 else src1.hi2)\n (if permutation % 2 = 0 then src1.lo1 else src1.hi3)\n (if (permutation / 2) % 2 = 0 then src2.lo0 else src2.hi2)\n (if (permutation / 2) % 2 = 0 then src2.lo1 else src2.hi3))\n else None\n\nlet eval_Shufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_sse2 (eval_Shufpd_raw permutation src1 src2)\nval ins_Shufpd (permutation:int) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Shufpd permutation)\n\nlet eval_VShufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Shufpd_raw permutation src1 src2)\nval ins_VShufpd (permutation:int) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VShufpd permutation)\n\nlet is_full_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x04050607 &&\n q.hi3 = 0x00010203\n\nlet is_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x00010203 &&\n q.lo1 = 0x04050607 &&\n q.hi2 = 0x08090A0B &&\n q.hi3 = 0x0C0D0E0F\n\nlet is_high_dup_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet is_lower_upper_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x04050607 &&\n q.lo1 = 0x00010203 &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet eval_Pshufb_raw (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if is_full_byte_reversal_mask src2 then\n Some (reverse_bytes_quad32 src1)\n else if is_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3))\n else if is_high_dup_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else if is_lower_upper_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else None\n\nlet eval_Pshufb (src1 src2:quad32) : option quad32 = check_ssse3 (eval_Pshufb_raw src1 src2)\nval ins_Pshufb : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pshufb\n\nlet eval_VPshufb (src1 src2:quad32) : option quad32 = check_avx (eval_Pshufb_raw src1 src2)\nval ins_VPshufb : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPshufb\n\nlet eval_Pshufd (permutation:nat8) (src:quad32) : option quad32 = check_sse2 (\n let bits:bits_of_byte = byte_to_twobits permutation in\n Some (Mkfour\n (select_word src bits.lo0)\n (select_word src bits.lo1)\n (select_word src bits.hi2)\n (select_word src bits.hi3)))\nval ins_Pshufd (permutation:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_Pshufd permutation)\n\nlet eval_Pcmpeqd (dst src:quad32) : option quad32 = check_sse2 (\n let eq_result (b:bool):nat32 = if b then 0xFFFFFFFF else 0 in\n Some (Mkfour\n (eq_result (src.lo0 = dst.lo0))\n (eq_result (src.lo1 = dst.lo1))\n (eq_result (src.hi2 = dst.hi2))\n (eq_result (src.hi3 = dst.hi3))))\nval ins_Pcmpeqd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pcmpeqd\n\nlet eval_Pextrq (index:nat8) (src:quad32) : option nat64 = check_sse4_1 (\n let src_two = four_to_two_two src in\n Some (two_to_nat 32 (two_select src_two (index % 2))))\nval ins_Pextrq (index:nat8) : instr_dep [out op64] [opXmm] PreserveFlags (eval_Pextrq index)\n\nlet eval_Pinsrd (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat32 dst (src % pow2_32) (index % 4)))\nval ins_Pinsrd (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrd index)\n\nlet eval_Pinsrq (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat64_def dst src (index % 2)))\nval ins_Pinsrq (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrq index)\n\nlet eval_Pslldq_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the two very special cases we need\n if count = 4 then Some (Mkfour 0 src.lo0 src.lo1 src.hi2)\n else if count = 8 then Some (Mkfour 0 0 src.lo0 src.lo1)\n else None\nlet eval_VPslldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Pslldq_raw count src)\nval ins_VPslldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPslldq count)\n\nlet eval_Psrldq_8_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the one very special case we need\n if count = 8 then Some (Mkfour src.hi2 src.hi3 0 0)\n else None\nlet eval_VPsrldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Psrldq_8_raw count src)\nval ins_VPsrldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPsrldq count)\n\nPartial implementation:\nlet eval_Pclmulqdq (imm: int) (src1 src2: quad32) : option quad32 =\n \n"
},
{
"role": "assistant",
"content": "let Mkfour a0 a1 a2 a3 = src1 in\nlet Mkfour b0 b1 b2 b3 = src2 in\nlet f x0 x1 y0 y1 =\n let x = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo x0 x1) in\n let y = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo y0 y1) in\n Vale.Math.Poly2.Bits_s.to_quad32 (Vale.Math.Poly2_s.mul x y)\nin\nif pclmulqdq_enabled\nthen\n match imm with\n | 0 -> Some (f a0 a1 b0 b1)\n | 1 -> Some (f a2 a3 b0 b1)\n | 16 -> Some (f a0 a1 b2 b3)\n | 17 -> Some (f a2 a3 b2 b3)\n | _ -> None\nelse None"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: src: Vale.X64.Machine_s.quad32 -> FStar.Pervasives.Native.option Vale.X64.Machine_s.quad32\nEffect: Prims.Tot\n\nContext:\nmodule Vale.X64.Instructions_s\nopen FStar.Mul\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Two_s\nopen Vale.Def.Words.Four_s\nopen Vale.Def.Types_s\nopen Vale.X64.Machine_s\nopen Vale.X64.Instruction_s\nopen Vale.X64.CPU_Features_s\nopen Vale.X64.CryptoInstructions_s\nopen FStar.Seq.Base\n\nlet eval_Mov64 (src:nat64) : option nat64 = Some src\nval ins_Mov64 : instr_dep [out op64] [op64] PreserveFlags eval_Mov64\n\nlet eval_MovBe64 (src:nat64) : option nat64 = if movbe_enabled then Some (reverse_bytes_nat64 src) else None\nval ins_MovBe64 : instr_dep [out op64] [op64] PreserveFlags eval_MovBe64\n\nlet eval_Bswap64 (dst:nat64) : option nat64 = Some (reverse_bytes_nat64 dst)\nval ins_Bswap64 : instr_dep [inOut op64] [] PreserveFlags eval_Bswap64\n\nlet eval_Cmovc64 (dst src:nat64) (carry:bool) : option nat64 = Some (if carry then src else dst)\nval ins_Cmovc64 : instr_dep [inOut op64] [op64; opFlagsCf] PreserveFlags eval_Cmovc64\n\nlet eval_Add64 (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src in Some (sum >= pow2_64, sum % pow2_64)\nval ins_Add64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Add64\n\nlet eval_AddLea64 (src1 src2:nat64) : option nat64 = Some ((src1 + src2) % pow2_64)\nval ins_AddLea64 :instr_dep [out op64] [op64; op64] PreserveFlags eval_AddLea64\n\nlet eval_AddCarry64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_carry then 1 else 0) in\n Some (sum >= pow2_64, sum % pow2_64)\nval ins_AddCarry64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_AddCarry64\n\nlet eval_Adcx64_Adox64 (old_flag:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_flag then 1 else 0) in\n if adx_enabled then Some (sum >= pow2_64, sum % pow2_64) else None\nval ins_Adcx64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\nval ins_Adox64 : instr_dep [inOut opFlagsOf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\n\nlet eval_Sub64 (dst src:nat64) : option (bool & nat64) =\n let diff = dst - src in Some (diff < 0, diff % pow2_64)\nval ins_Sub64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Sub64\n\nlet eval_Sbb64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let diff = dst - (src + (if old_carry then 1 else 0)) in\n Some (diff < 0, diff % pow2_64) // We specify cf, but underspecify everything else (via HavocFlags)\nval ins_Sbb64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_Sbb64\n\nlet eval_Mul64 (rax src:nat64) : option (nat64 & nat64) =\n Some (FStar.UInt.mul_div #64 rax src, FStar.UInt.mul_mod #64 rax src)\nval ins_Mul64 : instr_dep [out (one64Reg rRdx); inOut (one64Reg rRax)] [op64] HavocFlags eval_Mul64\n\nlet eval_Mulx64 (rdx src:nat64) : option (nat64 & nat64) =\n let hi = FStar.UInt.mul_div #64 rdx src in\n let lo = FStar.UInt.mul_mod #64 rdx src in\n if bmi2_enabled then Some (hi, lo) else None\nval ins_Mulx64 : instr_dep [out op64; out op64] [one64Reg rRdx; op64] PreserveFlags eval_Mulx64\n\nlet eval_IMul64 (dst src:nat64) : option nat64 =\n Some (FStar.UInt.mul_mod #64 dst src)\nval ins_IMul64 : instr_dep [inOut op64] [op64] HavocFlags eval_IMul64\n\nlet eval_And64 (dst src:nat64) : option nat64 = Some (iand dst src)\nval ins_And64 : instr_dep [inOut op64] [op64] HavocFlags eval_And64\n\nlet eval_Xor64 (dst src:nat64) : option (nat64 & (bool & bool)) =\n Some (Vale.Def.Types_s.ixor dst src, (false, false))\nval ins_Xor64 : instr_dep [inOut op64; out opFlagsCf; out opFlagsOf] [op64] HavocFlags eval_Xor64\n\nlet eval_Shr64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishr dst amt) else None\nval ins_Shr64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shr64\n\nlet eval_Shl64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishl dst amt) else None\nval ins_Shl64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shl64\n\nlet eval_Cpuid (rax rcx:nat64) : option (nat64 & (nat64 & (nat64 & nat64))) =\n Some (cpuid rRax rax rcx, (cpuid rRbx rax rcx, (cpuid rRcx rax rcx, cpuid rRdx rax rcx)))\nval ins_Cpuid :\n instr_dep\n [inOut (one64Reg rRax); out (one64Reg rRbx); inOut (one64Reg rRcx); out (one64Reg rRdx)]\n [] PreserveFlags eval_Cpuid\n\n// The XGETBV instruction requires that OSXSAVE (in CPUID) is enabled.\n// We underspecify XGETBV here to only support fetching XCR0, which\n// is supported on any processor supporting the XGETBV instruction\nlet eval_Xgetbv (rcx:nat64) : option (nat64 & nat64) =\n if osxsave_enabled && rcx = 0 then Some (xgetbv rRax rcx, xgetbv rRdx rcx) else None\nval ins_Xgetbv :\n instr_dep\n [out (one64Reg rRax); out (one64Reg rRdx)]\n [one64Reg rRcx] PreserveFlags eval_Xgetbv\n\nlet check_avx (#a:Type0) (x:option a) : option a =\n if avx_enabled then x else None\n\nlet check_sse2 (#a:Type0) (x:option a) : option a =\n if sse2_enabled then x else None\n\nlet check_ssse3 (#a:Type0) (x:option a) : option a =\n if ssse3_enabled then x else None\n\nlet check_sse4_1 (#a:Type0) (x:option a) : option a =\n if sse4_1_enabled then x else None\n\nlet eval_Movdqu (src:quad32) : option quad32 = check_sse2 (Some src)\nval ins_Movdqu : instr_dep [out opXmm] [opXmm] PreserveFlags eval_Movdqu\n\nlet eval_Pxor (dst src:quad32) : option quad32 = check_sse2 (Some (quad32_xor dst src))\nval ins_Pxor : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pxor\n\nlet eval_VPxor (src1 src2:quad32) : option quad32 = check_avx (Some (quad32_xor src1 src2))\nval ins_VPxor : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPxor\n\nlet eval_Pand (dst src:quad32) : option quad32 = check_sse2 (Some (four_map2 (fun di si -> iand di si) dst src))\nval ins_Pand : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pand\n\nlet eval_Paddd_raw (src1 src2:quad32) : option quad32 =\n Some (Mkfour\n ((src1.lo0 + src2.lo0) % pow2_32)\n ((src1.lo1 + src2.lo1) % pow2_32)\n ((src1.hi2 + src2.hi2) % pow2_32)\n ((src1.hi3 + src2.hi3) % pow2_32))\n\nlet eval_Paddd (src1 src2:quad32) : option quad32 = check_sse2 (eval_Paddd_raw src1 src2)\nval ins_Paddd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Paddd\n\nlet eval_VPaddd (src1 src2:quad32) : option quad32 = check_avx (eval_Paddd_raw src1 src2)\nval ins_VPaddd : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPaddd\n\nlet eval_Pslld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishl i amt) dst) else None)\nval ins_Pslld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Pslld amt)\n\nlet eval_Psrld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishr i amt) dst) else None)\nval ins_Psrld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrld amt)\n\nlet eval_Psrldq (amt:int) (dst:quad32) : option quad32 = check_sse2 (\n if 0 <= amt && amt < 16 then\n let src_bytes = le_quad32_to_bytes dst in\n let zero_pad = Seq.create amt 0 in\n let remaining_bytes = slice src_bytes amt (length src_bytes) in\n Some (le_bytes_to_quad32 (append zero_pad remaining_bytes))\n else None)\nval ins_Psrldq (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrldq amt)\n\nlet eval_Palignr_raw (amount:nat8) (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if amount = 4 then\n Some (Mkfour src2.lo1 src2.hi2 src2.hi3 src1.lo0)\n else if amount = 8 then\n Some (Mkfour src2.hi2 src2.hi3 src1.lo0 src1.lo1)\n else None\n\nlet eval_Palignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_ssse3 (eval_Palignr_raw amount src1 src2)\nval ins_Palignr (amount:nat8) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Palignr amount)\n\nlet eval_VPalignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Palignr_raw amount src1 src2)\nval ins_VPalignr (amount:nat8) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPalignr amount)\n\nlet eval_Shufpd_raw (permutation:int) (src1 src2:quad32) : option quad32 =\n if 0 <= permutation && permutation < 4 then\n Some (Mkfour\n (if permutation % 2 = 0 then src1.lo0 else src1.hi2)\n (if permutation % 2 = 0 then src1.lo1 else src1.hi3)\n (if (permutation / 2) % 2 = 0 then src2.lo0 else src2.hi2)\n (if (permutation / 2) % 2 = 0 then src2.lo1 else src2.hi3))\n else None\n\nlet eval_Shufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_sse2 (eval_Shufpd_raw permutation src1 src2)\nval ins_Shufpd (permutation:int) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Shufpd permutation)\n\nlet eval_VShufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Shufpd_raw permutation src1 src2)\nval ins_VShufpd (permutation:int) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VShufpd permutation)\n\nlet is_full_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x04050607 &&\n q.hi3 = 0x00010203\n\nlet is_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x00010203 &&\n q.lo1 = 0x04050607 &&\n q.hi2 = 0x08090A0B &&\n q.hi3 = 0x0C0D0E0F\n\nlet is_high_dup_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet is_lower_upper_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x04050607 &&\n q.lo1 = 0x00010203 &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet eval_Pshufb_raw (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if is_full_byte_reversal_mask src2 then\n Some (reverse_bytes_quad32 src1)\n else if is_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3))\n else if is_high_dup_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else if is_lower_upper_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else None\n\nlet eval_Pshufb (src1 src2:quad32) : option quad32 = check_ssse3 (eval_Pshufb_raw src1 src2)\nval ins_Pshufb : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pshufb\n\nlet eval_VPshufb (src1 src2:quad32) : option quad32 = check_avx (eval_Pshufb_raw src1 src2)\nval ins_VPshufb : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPshufb\n\nlet eval_Pshufd (permutation:nat8) (src:quad32) : option quad32 = check_sse2 (\n let bits:bits_of_byte = byte_to_twobits permutation in\n Some (Mkfour\n (select_word src bits.lo0)\n (select_word src bits.lo1)\n (select_word src bits.hi2)\n (select_word src bits.hi3)))\nval ins_Pshufd (permutation:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_Pshufd permutation)\n\nlet eval_Pcmpeqd (dst src:quad32) : option quad32 = check_sse2 (\n let eq_result (b:bool):nat32 = if b then 0xFFFFFFFF else 0 in\n Some (Mkfour\n (eq_result (src.lo0 = dst.lo0))\n (eq_result (src.lo1 = dst.lo1))\n (eq_result (src.hi2 = dst.hi2))\n (eq_result (src.hi3 = dst.hi3))))\nval ins_Pcmpeqd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pcmpeqd\n\nlet eval_Pextrq (index:nat8) (src:quad32) : option nat64 = check_sse4_1 (\n let src_two = four_to_two_two src in\n Some (two_to_nat 32 (two_select src_two (index % 2))))\nval ins_Pextrq (index:nat8) : instr_dep [out op64] [opXmm] PreserveFlags (eval_Pextrq index)\n\nlet eval_Pinsrd (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat32 dst (src % pow2_32) (index % 4)))\nval ins_Pinsrd (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrd index)\n\nlet eval_Pinsrq (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat64_def dst src (index % 2)))\nval ins_Pinsrq (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrq index)\n\nlet eval_Pslldq_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the two very special cases we need\n if count = 4 then Some (Mkfour 0 src.lo0 src.lo1 src.hi2)\n else if count = 8 then Some (Mkfour 0 0 src.lo0 src.lo1)\n else None\nlet eval_VPslldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Pslldq_raw count src)\nval ins_VPslldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPslldq count)\n\nlet eval_Psrldq_8_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the one very special case we need\n if count = 8 then Some (Mkfour src.hi2 src.hi3 0 0)\n else None\nlet eval_VPsrldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Psrldq_8_raw count src)\nval ins_VPsrldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPsrldq count)\n\nlet eval_Pclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n let Mkfour a0 a1 a2 a3 = src1 in\n let Mkfour b0 b1 b2 b3 = src2 in\n let f x0 x1 y0 y1 =\n let x = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo x0 x1) in\n let y = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo y0 y1) in\n Vale.Math.Poly2.Bits_s.to_quad32 (Vale.Math.Poly2_s.mul x y)\n in\n if pclmulqdq_enabled then\n match imm with\n | 0 -> Some (f a0 a1 b0 b1)\n | 1 -> Some (f a2 a3 b0 b1)\n | 16 -> Some (f a0 a1 b2 b3)\n | 17 -> Some (f a2 a3 b2 b3)\n | _ -> None\n else None\nval ins_Pclmulqdq (imm:int) : instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Pclmulqdq imm)\n\nlet eval_VPclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Pclmulqdq imm src1 src2)\nval ins_VPclmulqdq (imm:int) : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPclmulqdq imm)\n\nlet eval_AESNI_enc (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.mix_columns_LE (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1))) src2)\n else None\nval ins_AESNI_enc : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc\n\nlet eval_VAESNI_enc (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc src1 src2)\nval ins_VAESNI_enc : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc\n\nlet eval_AESNI_enc_last (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1)) src2)\n else None\nval ins_AESNI_enc_last : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc_last\n\nlet eval_VAESNI_enc_last (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc_last src1 src2)\nval ins_VAESNI_enc_last : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc_last\n\nlet eval_AESNI_dec (dst src:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.inv_mix_columns_LE (Vale.AES.AES_s.inv_sub_bytes (Vale.AES.AES_s.inv_shift_rows_LE dst))) src)\n else None\nval ins_AESNI_dec : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_dec\n\nlet eval_AESNI_dec_last (dst src:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.inv_sub_bytes (Vale.AES.AES_s.inv_shift_rows_LE dst)) src)\n else None\nval ins_AESNI_dec_last : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_dec_last\n\nPartial implementation:\nlet eval_AESNI_imc (src: quad32) : option quad32 =\n \n"
},
{
"role": "assistant",
"content": "if aesni_enabled then Some (Vale.AES.AES_s.inv_mix_columns_LE src) else None"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: _: Vale.X64.Machine_s.nat64 -> FStar.Pervasives.Native.option Prims.unit\nEffect: Prims.Tot\n\nContext:\nmodule Vale.X64.Instructions_s\nopen FStar.Mul\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Two_s\nopen Vale.Def.Words.Four_s\nopen Vale.Def.Types_s\nopen Vale.X64.Machine_s\nopen Vale.X64.Instruction_s\nopen Vale.X64.CPU_Features_s\nopen Vale.X64.CryptoInstructions_s\nopen FStar.Seq.Base\n\nlet eval_Mov64 (src:nat64) : option nat64 = Some src\nval ins_Mov64 : instr_dep [out op64] [op64] PreserveFlags eval_Mov64\n\nlet eval_MovBe64 (src:nat64) : option nat64 = if movbe_enabled then Some (reverse_bytes_nat64 src) else None\nval ins_MovBe64 : instr_dep [out op64] [op64] PreserveFlags eval_MovBe64\n\nlet eval_Bswap64 (dst:nat64) : option nat64 = Some (reverse_bytes_nat64 dst)\nval ins_Bswap64 : instr_dep [inOut op64] [] PreserveFlags eval_Bswap64\n\nlet eval_Cmovc64 (dst src:nat64) (carry:bool) : option nat64 = Some (if carry then src else dst)\nval ins_Cmovc64 : instr_dep [inOut op64] [op64; opFlagsCf] PreserveFlags eval_Cmovc64\n\nlet eval_Add64 (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src in Some (sum >= pow2_64, sum % pow2_64)\nval ins_Add64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Add64\n\nlet eval_AddLea64 (src1 src2:nat64) : option nat64 = Some ((src1 + src2) % pow2_64)\nval ins_AddLea64 :instr_dep [out op64] [op64; op64] PreserveFlags eval_AddLea64\n\nlet eval_AddCarry64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_carry then 1 else 0) in\n Some (sum >= pow2_64, sum % pow2_64)\nval ins_AddCarry64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_AddCarry64\n\nlet eval_Adcx64_Adox64 (old_flag:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_flag then 1 else 0) in\n if adx_enabled then Some (sum >= pow2_64, sum % pow2_64) else None\nval ins_Adcx64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\nval ins_Adox64 : instr_dep [inOut opFlagsOf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\n\nlet eval_Sub64 (dst src:nat64) : option (bool & nat64) =\n let diff = dst - src in Some (diff < 0, diff % pow2_64)\nval ins_Sub64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Sub64\n\nlet eval_Sbb64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let diff = dst - (src + (if old_carry then 1 else 0)) in\n Some (diff < 0, diff % pow2_64) // We specify cf, but underspecify everything else (via HavocFlags)\nval ins_Sbb64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_Sbb64\n\nlet eval_Mul64 (rax src:nat64) : option (nat64 & nat64) =\n Some (FStar.UInt.mul_div #64 rax src, FStar.UInt.mul_mod #64 rax src)\nval ins_Mul64 : instr_dep [out (one64Reg rRdx); inOut (one64Reg rRax)] [op64] HavocFlags eval_Mul64\n\nlet eval_Mulx64 (rdx src:nat64) : option (nat64 & nat64) =\n let hi = FStar.UInt.mul_div #64 rdx src in\n let lo = FStar.UInt.mul_mod #64 rdx src in\n if bmi2_enabled then Some (hi, lo) else None\nval ins_Mulx64 : instr_dep [out op64; out op64] [one64Reg rRdx; op64] PreserveFlags eval_Mulx64\n\nlet eval_IMul64 (dst src:nat64) : option nat64 =\n Some (FStar.UInt.mul_mod #64 dst src)\nval ins_IMul64 : instr_dep [inOut op64] [op64] HavocFlags eval_IMul64\n\nlet eval_And64 (dst src:nat64) : option nat64 = Some (iand dst src)\nval ins_And64 : instr_dep [inOut op64] [op64] HavocFlags eval_And64\n\nlet eval_Xor64 (dst src:nat64) : option (nat64 & (bool & bool)) =\n Some (Vale.Def.Types_s.ixor dst src, (false, false))\nval ins_Xor64 : instr_dep [inOut op64; out opFlagsCf; out opFlagsOf] [op64] HavocFlags eval_Xor64\n\nlet eval_Shr64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishr dst amt) else None\nval ins_Shr64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shr64\n\nlet eval_Shl64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishl dst amt) else None\nval ins_Shl64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shl64\n\nlet eval_Cpuid (rax rcx:nat64) : option (nat64 & (nat64 & (nat64 & nat64))) =\n Some (cpuid rRax rax rcx, (cpuid rRbx rax rcx, (cpuid rRcx rax rcx, cpuid rRdx rax rcx)))\nval ins_Cpuid :\n instr_dep\n [inOut (one64Reg rRax); out (one64Reg rRbx); inOut (one64Reg rRcx); out (one64Reg rRdx)]\n [] PreserveFlags eval_Cpuid\n\n// The XGETBV instruction requires that OSXSAVE (in CPUID) is enabled.\n// We underspecify XGETBV here to only support fetching XCR0, which\n// is supported on any processor supporting the XGETBV instruction\nlet eval_Xgetbv (rcx:nat64) : option (nat64 & nat64) =\n if osxsave_enabled && rcx = 0 then Some (xgetbv rRax rcx, xgetbv rRdx rcx) else None\nval ins_Xgetbv :\n instr_dep\n [out (one64Reg rRax); out (one64Reg rRdx)]\n [one64Reg rRcx] PreserveFlags eval_Xgetbv\n\nlet check_avx (#a:Type0) (x:option a) : option a =\n if avx_enabled then x else None\n\nlet check_sse2 (#a:Type0) (x:option a) : option a =\n if sse2_enabled then x else None\n\nlet check_ssse3 (#a:Type0) (x:option a) : option a =\n if ssse3_enabled then x else None\n\nlet check_sse4_1 (#a:Type0) (x:option a) : option a =\n if sse4_1_enabled then x else None\n\nlet eval_Movdqu (src:quad32) : option quad32 = check_sse2 (Some src)\nval ins_Movdqu : instr_dep [out opXmm] [opXmm] PreserveFlags eval_Movdqu\n\nlet eval_Pxor (dst src:quad32) : option quad32 = check_sse2 (Some (quad32_xor dst src))\nval ins_Pxor : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pxor\n\nlet eval_VPxor (src1 src2:quad32) : option quad32 = check_avx (Some (quad32_xor src1 src2))\nval ins_VPxor : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPxor\n\nlet eval_Pand (dst src:quad32) : option quad32 = check_sse2 (Some (four_map2 (fun di si -> iand di si) dst src))\nval ins_Pand : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pand\n\nlet eval_Paddd_raw (src1 src2:quad32) : option quad32 =\n Some (Mkfour\n ((src1.lo0 + src2.lo0) % pow2_32)\n ((src1.lo1 + src2.lo1) % pow2_32)\n ((src1.hi2 + src2.hi2) % pow2_32)\n ((src1.hi3 + src2.hi3) % pow2_32))\n\nlet eval_Paddd (src1 src2:quad32) : option quad32 = check_sse2 (eval_Paddd_raw src1 src2)\nval ins_Paddd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Paddd\n\nlet eval_VPaddd (src1 src2:quad32) : option quad32 = check_avx (eval_Paddd_raw src1 src2)\nval ins_VPaddd : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPaddd\n\nlet eval_Pslld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishl i amt) dst) else None)\nval ins_Pslld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Pslld amt)\n\nlet eval_Psrld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishr i amt) dst) else None)\nval ins_Psrld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrld amt)\n\nlet eval_Psrldq (amt:int) (dst:quad32) : option quad32 = check_sse2 (\n if 0 <= amt && amt < 16 then\n let src_bytes = le_quad32_to_bytes dst in\n let zero_pad = Seq.create amt 0 in\n let remaining_bytes = slice src_bytes amt (length src_bytes) in\n Some (le_bytes_to_quad32 (append zero_pad remaining_bytes))\n else None)\nval ins_Psrldq (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrldq amt)\n\nlet eval_Palignr_raw (amount:nat8) (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if amount = 4 then\n Some (Mkfour src2.lo1 src2.hi2 src2.hi3 src1.lo0)\n else if amount = 8 then\n Some (Mkfour src2.hi2 src2.hi3 src1.lo0 src1.lo1)\n else None\n\nlet eval_Palignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_ssse3 (eval_Palignr_raw amount src1 src2)\nval ins_Palignr (amount:nat8) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Palignr amount)\n\nlet eval_VPalignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Palignr_raw amount src1 src2)\nval ins_VPalignr (amount:nat8) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPalignr amount)\n\nlet eval_Shufpd_raw (permutation:int) (src1 src2:quad32) : option quad32 =\n if 0 <= permutation && permutation < 4 then\n Some (Mkfour\n (if permutation % 2 = 0 then src1.lo0 else src1.hi2)\n (if permutation % 2 = 0 then src1.lo1 else src1.hi3)\n (if (permutation / 2) % 2 = 0 then src2.lo0 else src2.hi2)\n (if (permutation / 2) % 2 = 0 then src2.lo1 else src2.hi3))\n else None\n\nlet eval_Shufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_sse2 (eval_Shufpd_raw permutation src1 src2)\nval ins_Shufpd (permutation:int) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Shufpd permutation)\n\nlet eval_VShufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Shufpd_raw permutation src1 src2)\nval ins_VShufpd (permutation:int) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VShufpd permutation)\n\nlet is_full_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x04050607 &&\n q.hi3 = 0x00010203\n\nlet is_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x00010203 &&\n q.lo1 = 0x04050607 &&\n q.hi2 = 0x08090A0B &&\n q.hi3 = 0x0C0D0E0F\n\nlet is_high_dup_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet is_lower_upper_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x04050607 &&\n q.lo1 = 0x00010203 &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet eval_Pshufb_raw (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if is_full_byte_reversal_mask src2 then\n Some (reverse_bytes_quad32 src1)\n else if is_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3))\n else if is_high_dup_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else if is_lower_upper_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else None\n\nlet eval_Pshufb (src1 src2:quad32) : option quad32 = check_ssse3 (eval_Pshufb_raw src1 src2)\nval ins_Pshufb : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pshufb\n\nlet eval_VPshufb (src1 src2:quad32) : option quad32 = check_avx (eval_Pshufb_raw src1 src2)\nval ins_VPshufb : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPshufb\n\nlet eval_Pshufd (permutation:nat8) (src:quad32) : option quad32 = check_sse2 (\n let bits:bits_of_byte = byte_to_twobits permutation in\n Some (Mkfour\n (select_word src bits.lo0)\n (select_word src bits.lo1)\n (select_word src bits.hi2)\n (select_word src bits.hi3)))\nval ins_Pshufd (permutation:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_Pshufd permutation)\n\nlet eval_Pcmpeqd (dst src:quad32) : option quad32 = check_sse2 (\n let eq_result (b:bool):nat32 = if b then 0xFFFFFFFF else 0 in\n Some (Mkfour\n (eq_result (src.lo0 = dst.lo0))\n (eq_result (src.lo1 = dst.lo1))\n (eq_result (src.hi2 = dst.hi2))\n (eq_result (src.hi3 = dst.hi3))))\nval ins_Pcmpeqd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pcmpeqd\n\nlet eval_Pextrq (index:nat8) (src:quad32) : option nat64 = check_sse4_1 (\n let src_two = four_to_two_two src in\n Some (two_to_nat 32 (two_select src_two (index % 2))))\nval ins_Pextrq (index:nat8) : instr_dep [out op64] [opXmm] PreserveFlags (eval_Pextrq index)\n\nlet eval_Pinsrd (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat32 dst (src % pow2_32) (index % 4)))\nval ins_Pinsrd (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrd index)\n\nlet eval_Pinsrq (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat64_def dst src (index % 2)))\nval ins_Pinsrq (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrq index)\n\nlet eval_Pslldq_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the two very special cases we need\n if count = 4 then Some (Mkfour 0 src.lo0 src.lo1 src.hi2)\n else if count = 8 then Some (Mkfour 0 0 src.lo0 src.lo1)\n else None\nlet eval_VPslldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Pslldq_raw count src)\nval ins_VPslldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPslldq count)\n\nlet eval_Psrldq_8_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the one very special case we need\n if count = 8 then Some (Mkfour src.hi2 src.hi3 0 0)\n else None\nlet eval_VPsrldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Psrldq_8_raw count src)\nval ins_VPsrldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPsrldq count)\n\nlet eval_Pclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n let Mkfour a0 a1 a2 a3 = src1 in\n let Mkfour b0 b1 b2 b3 = src2 in\n let f x0 x1 y0 y1 =\n let x = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo x0 x1) in\n let y = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo y0 y1) in\n Vale.Math.Poly2.Bits_s.to_quad32 (Vale.Math.Poly2_s.mul x y)\n in\n if pclmulqdq_enabled then\n match imm with\n | 0 -> Some (f a0 a1 b0 b1)\n | 1 -> Some (f a2 a3 b0 b1)\n | 16 -> Some (f a0 a1 b2 b3)\n | 17 -> Some (f a2 a3 b2 b3)\n | _ -> None\n else None\nval ins_Pclmulqdq (imm:int) : instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Pclmulqdq imm)\n\nlet eval_VPclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Pclmulqdq imm src1 src2)\nval ins_VPclmulqdq (imm:int) : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPclmulqdq imm)\n\nlet eval_AESNI_enc (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.mix_columns_LE (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1))) src2)\n else None\nval ins_AESNI_enc : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc\n\nlet eval_VAESNI_enc (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc src1 src2)\nval ins_VAESNI_enc : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc\n\nlet eval_AESNI_enc_last (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1)) src2)\n else None\nval ins_AESNI_enc_last : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc_last\n\nlet eval_VAESNI_enc_last (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc_last src1 src2)\nval ins_VAESNI_enc_last : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc_last\n\nlet eval_AESNI_dec (dst src:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.inv_mix_columns_LE (Vale.AES.AES_s.inv_sub_bytes (Vale.AES.AES_s.inv_shift_rows_LE dst))) src)\n else None\nval ins_AESNI_dec : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_dec\n\nlet eval_AESNI_dec_last (dst src:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.inv_sub_bytes (Vale.AES.AES_s.inv_shift_rows_LE dst)) src)\n else None\nval ins_AESNI_dec_last : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_dec_last\n\nlet eval_AESNI_imc (src:quad32) : option quad32 =\n if aesni_enabled then Some (Vale.AES.AES_s.inv_mix_columns_LE src) else None\nval ins_AESNI_imc : instr_dep [out opXmm] [opXmm] PreserveFlags eval_AESNI_imc\n\nlet eval_AESNI_keygen_assist (imm:nat8) (src:quad32) : option quad32 =\n if aesni_enabled then\n Some (Mkfour\n (Vale.AES.AES_s.sub_word src.lo1)\n (ixor (Vale.AES.AES_s.rot_word_LE (Vale.AES.AES_s.sub_word src.lo1)) imm)\n (Vale.AES.AES_s.sub_word src.hi3)\n (ixor (Vale.AES.AES_s.rot_word_LE (Vale.AES.AES_s.sub_word src.hi3)) imm))\n else None\nval ins_AESNI_keygen_assist (imm:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_AESNI_keygen_assist imm)\n\nlet eval_SHA256_rnds2 (src1 src2 wk:quad32) : option quad32 =\n if sha_enabled then Some (sha256_rnds2_spec src1 src2 wk) else None\nval ins_SHA256_rnds2 :\n instr_dep [inOut opXmm] [opXmm; oneXmm (OReg 0)] PreserveFlags eval_SHA256_rnds2\n\nlet eval_SHA256_msg1 (src1 src2:quad32) : option quad32 =\n if sha_enabled then Some (sha256_msg1_spec src1 src2) else None\nval ins_SHA256_msg1 : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_SHA256_msg1\n\nlet eval_SHA256_msg2 (src1 src2:quad32) : option quad32 =\n if sha_enabled then Some (sha256_msg2_spec src1 src2) else None\nval ins_SHA256_msg2 : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_SHA256_msg2\n\nlet eval_Ghost : option unit = Some ()\nval ins_Ghost : instr_dep [] [] PreserveFlags eval_Ghost\n\nlet eval_Comment : option unit = Some ()\nval ins_Comment (_:string) : instr_dep [] [] PreserveFlags eval_Comment\n\nlet eval_LargeComment : option unit = Some ()\nval ins_LargeComment (_:string) : instr_dep [] [] PreserveFlags eval_LargeComment\n\nlet eval_Newline : option unit = Some ()\nval ins_Newline : instr_dep [] [] PreserveFlags eval_Newline\n\nlet eval_Space : option unit = Some ()\nval ins_Space (_:nat) : instr_dep [] [] PreserveFlags eval_Space\n\nPartial implementation:\nlet eval_Prefetchnta (_: nat64) : option unit =\n \n"
},
{
"role": "assistant",
"content": "Some ()"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: src1: Vale.X64.Machine_s.quad32 -> src2: Vale.X64.Machine_s.quad32\n -> FStar.Pervasives.Native.option Vale.X64.Machine_s.quad32\nEffect: Prims.Tot\n\nContext:\nmodule Vale.X64.Instructions_s\nopen FStar.Mul\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Two_s\nopen Vale.Def.Words.Four_s\nopen Vale.Def.Types_s\nopen Vale.X64.Machine_s\nopen Vale.X64.Instruction_s\nopen Vale.X64.CPU_Features_s\nopen Vale.X64.CryptoInstructions_s\nopen FStar.Seq.Base\n\nlet eval_Mov64 (src:nat64) : option nat64 = Some src\nval ins_Mov64 : instr_dep [out op64] [op64] PreserveFlags eval_Mov64\n\nlet eval_MovBe64 (src:nat64) : option nat64 = if movbe_enabled then Some (reverse_bytes_nat64 src) else None\nval ins_MovBe64 : instr_dep [out op64] [op64] PreserveFlags eval_MovBe64\n\nlet eval_Bswap64 (dst:nat64) : option nat64 = Some (reverse_bytes_nat64 dst)\nval ins_Bswap64 : instr_dep [inOut op64] [] PreserveFlags eval_Bswap64\n\nlet eval_Cmovc64 (dst src:nat64) (carry:bool) : option nat64 = Some (if carry then src else dst)\nval ins_Cmovc64 : instr_dep [inOut op64] [op64; opFlagsCf] PreserveFlags eval_Cmovc64\n\nlet eval_Add64 (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src in Some (sum >= pow2_64, sum % pow2_64)\nval ins_Add64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Add64\n\nlet eval_AddLea64 (src1 src2:nat64) : option nat64 = Some ((src1 + src2) % pow2_64)\nval ins_AddLea64 :instr_dep [out op64] [op64; op64] PreserveFlags eval_AddLea64\n\nlet eval_AddCarry64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_carry then 1 else 0) in\n Some (sum >= pow2_64, sum % pow2_64)\nval ins_AddCarry64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_AddCarry64\n\nlet eval_Adcx64_Adox64 (old_flag:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_flag then 1 else 0) in\n if adx_enabled then Some (sum >= pow2_64, sum % pow2_64) else None\nval ins_Adcx64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\nval ins_Adox64 : instr_dep [inOut opFlagsOf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\n\nlet eval_Sub64 (dst src:nat64) : option (bool & nat64) =\n let diff = dst - src in Some (diff < 0, diff % pow2_64)\nval ins_Sub64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Sub64\n\nlet eval_Sbb64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let diff = dst - (src + (if old_carry then 1 else 0)) in\n Some (diff < 0, diff % pow2_64) // We specify cf, but underspecify everything else (via HavocFlags)\nval ins_Sbb64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_Sbb64\n\nlet eval_Mul64 (rax src:nat64) : option (nat64 & nat64) =\n Some (FStar.UInt.mul_div #64 rax src, FStar.UInt.mul_mod #64 rax src)\nval ins_Mul64 : instr_dep [out (one64Reg rRdx); inOut (one64Reg rRax)] [op64] HavocFlags eval_Mul64\n\nlet eval_Mulx64 (rdx src:nat64) : option (nat64 & nat64) =\n let hi = FStar.UInt.mul_div #64 rdx src in\n let lo = FStar.UInt.mul_mod #64 rdx src in\n if bmi2_enabled then Some (hi, lo) else None\nval ins_Mulx64 : instr_dep [out op64; out op64] [one64Reg rRdx; op64] PreserveFlags eval_Mulx64\n\nlet eval_IMul64 (dst src:nat64) : option nat64 =\n Some (FStar.UInt.mul_mod #64 dst src)\nval ins_IMul64 : instr_dep [inOut op64] [op64] HavocFlags eval_IMul64\n\nlet eval_And64 (dst src:nat64) : option nat64 = Some (iand dst src)\nval ins_And64 : instr_dep [inOut op64] [op64] HavocFlags eval_And64\n\nlet eval_Xor64 (dst src:nat64) : option (nat64 & (bool & bool)) =\n Some (Vale.Def.Types_s.ixor dst src, (false, false))\nval ins_Xor64 : instr_dep [inOut op64; out opFlagsCf; out opFlagsOf] [op64] HavocFlags eval_Xor64\n\nlet eval_Shr64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishr dst amt) else None\nval ins_Shr64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shr64\n\nlet eval_Shl64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishl dst amt) else None\nval ins_Shl64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shl64\n\nlet eval_Cpuid (rax rcx:nat64) : option (nat64 & (nat64 & (nat64 & nat64))) =\n Some (cpuid rRax rax rcx, (cpuid rRbx rax rcx, (cpuid rRcx rax rcx, cpuid rRdx rax rcx)))\nval ins_Cpuid :\n instr_dep\n [inOut (one64Reg rRax); out (one64Reg rRbx); inOut (one64Reg rRcx); out (one64Reg rRdx)]\n [] PreserveFlags eval_Cpuid\n\n// The XGETBV instruction requires that OSXSAVE (in CPUID) is enabled.\n// We underspecify XGETBV here to only support fetching XCR0, which\n// is supported on any processor supporting the XGETBV instruction\nlet eval_Xgetbv (rcx:nat64) : option (nat64 & nat64) =\n if osxsave_enabled && rcx = 0 then Some (xgetbv rRax rcx, xgetbv rRdx rcx) else None\nval ins_Xgetbv :\n instr_dep\n [out (one64Reg rRax); out (one64Reg rRdx)]\n [one64Reg rRcx] PreserveFlags eval_Xgetbv\n\nlet check_avx (#a:Type0) (x:option a) : option a =\n if avx_enabled then x else None\n\nlet check_sse2 (#a:Type0) (x:option a) : option a =\n if sse2_enabled then x else None\n\nlet check_ssse3 (#a:Type0) (x:option a) : option a =\n if ssse3_enabled then x else None\n\nlet check_sse4_1 (#a:Type0) (x:option a) : option a =\n if sse4_1_enabled then x else None\n\nlet eval_Movdqu (src:quad32) : option quad32 = check_sse2 (Some src)\nval ins_Movdqu : instr_dep [out opXmm] [opXmm] PreserveFlags eval_Movdqu\n\nlet eval_Pxor (dst src:quad32) : option quad32 = check_sse2 (Some (quad32_xor dst src))\nval ins_Pxor : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pxor\n\nlet eval_VPxor (src1 src2:quad32) : option quad32 = check_avx (Some (quad32_xor src1 src2))\nval ins_VPxor : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPxor\n\nlet eval_Pand (dst src:quad32) : option quad32 = check_sse2 (Some (four_map2 (fun di si -> iand di si) dst src))\nval ins_Pand : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pand\n\nlet eval_Paddd_raw (src1 src2:quad32) : option quad32 =\n Some (Mkfour\n ((src1.lo0 + src2.lo0) % pow2_32)\n ((src1.lo1 + src2.lo1) % pow2_32)\n ((src1.hi2 + src2.hi2) % pow2_32)\n ((src1.hi3 + src2.hi3) % pow2_32))\n\nlet eval_Paddd (src1 src2:quad32) : option quad32 = check_sse2 (eval_Paddd_raw src1 src2)\nval ins_Paddd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Paddd\n\nlet eval_VPaddd (src1 src2:quad32) : option quad32 = check_avx (eval_Paddd_raw src1 src2)\nval ins_VPaddd : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPaddd\n\nlet eval_Pslld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishl i amt) dst) else None)\nval ins_Pslld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Pslld amt)\n\nlet eval_Psrld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishr i amt) dst) else None)\nval ins_Psrld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrld amt)\n\nlet eval_Psrldq (amt:int) (dst:quad32) : option quad32 = check_sse2 (\n if 0 <= amt && amt < 16 then\n let src_bytes = le_quad32_to_bytes dst in\n let zero_pad = Seq.create amt 0 in\n let remaining_bytes = slice src_bytes amt (length src_bytes) in\n Some (le_bytes_to_quad32 (append zero_pad remaining_bytes))\n else None)\nval ins_Psrldq (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrldq amt)\n\nlet eval_Palignr_raw (amount:nat8) (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if amount = 4 then\n Some (Mkfour src2.lo1 src2.hi2 src2.hi3 src1.lo0)\n else if amount = 8 then\n Some (Mkfour src2.hi2 src2.hi3 src1.lo0 src1.lo1)\n else None\n\nlet eval_Palignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_ssse3 (eval_Palignr_raw amount src1 src2)\nval ins_Palignr (amount:nat8) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Palignr amount)\n\nlet eval_VPalignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Palignr_raw amount src1 src2)\nval ins_VPalignr (amount:nat8) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPalignr amount)\n\nlet eval_Shufpd_raw (permutation:int) (src1 src2:quad32) : option quad32 =\n if 0 <= permutation && permutation < 4 then\n Some (Mkfour\n (if permutation % 2 = 0 then src1.lo0 else src1.hi2)\n (if permutation % 2 = 0 then src1.lo1 else src1.hi3)\n (if (permutation / 2) % 2 = 0 then src2.lo0 else src2.hi2)\n (if (permutation / 2) % 2 = 0 then src2.lo1 else src2.hi3))\n else None\n\nlet eval_Shufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_sse2 (eval_Shufpd_raw permutation src1 src2)\nval ins_Shufpd (permutation:int) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Shufpd permutation)\n\nlet eval_VShufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Shufpd_raw permutation src1 src2)\nval ins_VShufpd (permutation:int) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VShufpd permutation)\n\nlet is_full_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x04050607 &&\n q.hi3 = 0x00010203\n\nlet is_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x00010203 &&\n q.lo1 = 0x04050607 &&\n q.hi2 = 0x08090A0B &&\n q.hi3 = 0x0C0D0E0F\n\nlet is_high_dup_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet is_lower_upper_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x04050607 &&\n q.lo1 = 0x00010203 &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet eval_Pshufb_raw (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if is_full_byte_reversal_mask src2 then\n Some (reverse_bytes_quad32 src1)\n else if is_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3))\n else if is_high_dup_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else if is_lower_upper_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else None\n\nlet eval_Pshufb (src1 src2:quad32) : option quad32 = check_ssse3 (eval_Pshufb_raw src1 src2)\nval ins_Pshufb : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pshufb\n\nlet eval_VPshufb (src1 src2:quad32) : option quad32 = check_avx (eval_Pshufb_raw src1 src2)\nval ins_VPshufb : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPshufb\n\nlet eval_Pshufd (permutation:nat8) (src:quad32) : option quad32 = check_sse2 (\n let bits:bits_of_byte = byte_to_twobits permutation in\n Some (Mkfour\n (select_word src bits.lo0)\n (select_word src bits.lo1)\n (select_word src bits.hi2)\n (select_word src bits.hi3)))\nval ins_Pshufd (permutation:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_Pshufd permutation)\n\nlet eval_Pcmpeqd (dst src:quad32) : option quad32 = check_sse2 (\n let eq_result (b:bool):nat32 = if b then 0xFFFFFFFF else 0 in\n Some (Mkfour\n (eq_result (src.lo0 = dst.lo0))\n (eq_result (src.lo1 = dst.lo1))\n (eq_result (src.hi2 = dst.hi2))\n (eq_result (src.hi3 = dst.hi3))))\nval ins_Pcmpeqd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pcmpeqd\n\nlet eval_Pextrq (index:nat8) (src:quad32) : option nat64 = check_sse4_1 (\n let src_two = four_to_two_two src in\n Some (two_to_nat 32 (two_select src_two (index % 2))))\nval ins_Pextrq (index:nat8) : instr_dep [out op64] [opXmm] PreserveFlags (eval_Pextrq index)\n\nlet eval_Pinsrd (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat32 dst (src % pow2_32) (index % 4)))\nval ins_Pinsrd (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrd index)\n\nlet eval_Pinsrq (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat64_def dst src (index % 2)))\nval ins_Pinsrq (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrq index)\n\nlet eval_Pslldq_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the two very special cases we need\n if count = 4 then Some (Mkfour 0 src.lo0 src.lo1 src.hi2)\n else if count = 8 then Some (Mkfour 0 0 src.lo0 src.lo1)\n else None\nlet eval_VPslldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Pslldq_raw count src)\nval ins_VPslldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPslldq count)\n\nlet eval_Psrldq_8_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the one very special case we need\n if count = 8 then Some (Mkfour src.hi2 src.hi3 0 0)\n else None\nlet eval_VPsrldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Psrldq_8_raw count src)\nval ins_VPsrldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPsrldq count)\n\nlet eval_Pclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n let Mkfour a0 a1 a2 a3 = src1 in\n let Mkfour b0 b1 b2 b3 = src2 in\n let f x0 x1 y0 y1 =\n let x = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo x0 x1) in\n let y = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo y0 y1) in\n Vale.Math.Poly2.Bits_s.to_quad32 (Vale.Math.Poly2_s.mul x y)\n in\n if pclmulqdq_enabled then\n match imm with\n | 0 -> Some (f a0 a1 b0 b1)\n | 1 -> Some (f a2 a3 b0 b1)\n | 16 -> Some (f a0 a1 b2 b3)\n | 17 -> Some (f a2 a3 b2 b3)\n | _ -> None\n else None\nval ins_Pclmulqdq (imm:int) : instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Pclmulqdq imm)\n\nlet eval_VPclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Pclmulqdq imm src1 src2)\nval ins_VPclmulqdq (imm:int) : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPclmulqdq imm)\n\nlet eval_AESNI_enc (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.mix_columns_LE (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1))) src2)\n else None\nval ins_AESNI_enc : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc\n\nlet eval_VAESNI_enc (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc src1 src2)\nval ins_VAESNI_enc : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc\n\nlet eval_AESNI_enc_last (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1)) src2)\n else None\nval ins_AESNI_enc_last : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc_last\n\nlet eval_VAESNI_enc_last (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc_last src1 src2)\nval ins_VAESNI_enc_last : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc_last\n\nlet eval_AESNI_dec (dst src:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.inv_mix_columns_LE (Vale.AES.AES_s.inv_sub_bytes (Vale.AES.AES_s.inv_shift_rows_LE dst))) src)\n else None\nval ins_AESNI_dec : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_dec\n\nlet eval_AESNI_dec_last (dst src:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.inv_sub_bytes (Vale.AES.AES_s.inv_shift_rows_LE dst)) src)\n else None\nval ins_AESNI_dec_last : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_dec_last\n\nlet eval_AESNI_imc (src:quad32) : option quad32 =\n if aesni_enabled then Some (Vale.AES.AES_s.inv_mix_columns_LE src) else None\nval ins_AESNI_imc : instr_dep [out opXmm] [opXmm] PreserveFlags eval_AESNI_imc\n\nlet eval_AESNI_keygen_assist (imm:nat8) (src:quad32) : option quad32 =\n if aesni_enabled then\n Some (Mkfour\n (Vale.AES.AES_s.sub_word src.lo1)\n (ixor (Vale.AES.AES_s.rot_word_LE (Vale.AES.AES_s.sub_word src.lo1)) imm)\n (Vale.AES.AES_s.sub_word src.hi3)\n (ixor (Vale.AES.AES_s.rot_word_LE (Vale.AES.AES_s.sub_word src.hi3)) imm))\n else None\nval ins_AESNI_keygen_assist (imm:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_AESNI_keygen_assist imm)\n\nlet eval_SHA256_rnds2 (src1 src2 wk:quad32) : option quad32 =\n if sha_enabled then Some (sha256_rnds2_spec src1 src2 wk) else None\nval ins_SHA256_rnds2 :\n instr_dep [inOut opXmm] [opXmm; oneXmm (OReg 0)] PreserveFlags eval_SHA256_rnds2\n\nPartial implementation:\nlet eval_SHA256_msg1 (src1 src2: quad32) : option quad32 =\n \n"
},
{
"role": "assistant",
"content": "if sha_enabled then Some (sha256_msg1_spec src1 src2) else None"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: dst: Vale.X64.Machine_s.quad32 -> src: Vale.X64.Machine_s.quad32\n -> FStar.Pervasives.Native.option Vale.X64.Machine_s.quad32\nEffect: Prims.Tot\n\nContext:\nmodule Vale.X64.Instructions_s\nopen FStar.Mul\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Two_s\nopen Vale.Def.Words.Four_s\nopen Vale.Def.Types_s\nopen Vale.X64.Machine_s\nopen Vale.X64.Instruction_s\nopen Vale.X64.CPU_Features_s\nopen Vale.X64.CryptoInstructions_s\nopen FStar.Seq.Base\n\nlet eval_Mov64 (src:nat64) : option nat64 = Some src\nval ins_Mov64 : instr_dep [out op64] [op64] PreserveFlags eval_Mov64\n\nlet eval_MovBe64 (src:nat64) : option nat64 = if movbe_enabled then Some (reverse_bytes_nat64 src) else None\nval ins_MovBe64 : instr_dep [out op64] [op64] PreserveFlags eval_MovBe64\n\nlet eval_Bswap64 (dst:nat64) : option nat64 = Some (reverse_bytes_nat64 dst)\nval ins_Bswap64 : instr_dep [inOut op64] [] PreserveFlags eval_Bswap64\n\nlet eval_Cmovc64 (dst src:nat64) (carry:bool) : option nat64 = Some (if carry then src else dst)\nval ins_Cmovc64 : instr_dep [inOut op64] [op64; opFlagsCf] PreserveFlags eval_Cmovc64\n\nlet eval_Add64 (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src in Some (sum >= pow2_64, sum % pow2_64)\nval ins_Add64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Add64\n\nlet eval_AddLea64 (src1 src2:nat64) : option nat64 = Some ((src1 + src2) % pow2_64)\nval ins_AddLea64 :instr_dep [out op64] [op64; op64] PreserveFlags eval_AddLea64\n\nlet eval_AddCarry64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_carry then 1 else 0) in\n Some (sum >= pow2_64, sum % pow2_64)\nval ins_AddCarry64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_AddCarry64\n\nlet eval_Adcx64_Adox64 (old_flag:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_flag then 1 else 0) in\n if adx_enabled then Some (sum >= pow2_64, sum % pow2_64) else None\nval ins_Adcx64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\nval ins_Adox64 : instr_dep [inOut opFlagsOf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\n\nlet eval_Sub64 (dst src:nat64) : option (bool & nat64) =\n let diff = dst - src in Some (diff < 0, diff % pow2_64)\nval ins_Sub64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Sub64\n\nlet eval_Sbb64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let diff = dst - (src + (if old_carry then 1 else 0)) in\n Some (diff < 0, diff % pow2_64) // We specify cf, but underspecify everything else (via HavocFlags)\nval ins_Sbb64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_Sbb64\n\nlet eval_Mul64 (rax src:nat64) : option (nat64 & nat64) =\n Some (FStar.UInt.mul_div #64 rax src, FStar.UInt.mul_mod #64 rax src)\nval ins_Mul64 : instr_dep [out (one64Reg rRdx); inOut (one64Reg rRax)] [op64] HavocFlags eval_Mul64\n\nlet eval_Mulx64 (rdx src:nat64) : option (nat64 & nat64) =\n let hi = FStar.UInt.mul_div #64 rdx src in\n let lo = FStar.UInt.mul_mod #64 rdx src in\n if bmi2_enabled then Some (hi, lo) else None\nval ins_Mulx64 : instr_dep [out op64; out op64] [one64Reg rRdx; op64] PreserveFlags eval_Mulx64\n\nlet eval_IMul64 (dst src:nat64) : option nat64 =\n Some (FStar.UInt.mul_mod #64 dst src)\nval ins_IMul64 : instr_dep [inOut op64] [op64] HavocFlags eval_IMul64\n\nlet eval_And64 (dst src:nat64) : option nat64 = Some (iand dst src)\nval ins_And64 : instr_dep [inOut op64] [op64] HavocFlags eval_And64\n\nlet eval_Xor64 (dst src:nat64) : option (nat64 & (bool & bool)) =\n Some (Vale.Def.Types_s.ixor dst src, (false, false))\nval ins_Xor64 : instr_dep [inOut op64; out opFlagsCf; out opFlagsOf] [op64] HavocFlags eval_Xor64\n\nlet eval_Shr64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishr dst amt) else None\nval ins_Shr64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shr64\n\nlet eval_Shl64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishl dst amt) else None\nval ins_Shl64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shl64\n\nlet eval_Cpuid (rax rcx:nat64) : option (nat64 & (nat64 & (nat64 & nat64))) =\n Some (cpuid rRax rax rcx, (cpuid rRbx rax rcx, (cpuid rRcx rax rcx, cpuid rRdx rax rcx)))\nval ins_Cpuid :\n instr_dep\n [inOut (one64Reg rRax); out (one64Reg rRbx); inOut (one64Reg rRcx); out (one64Reg rRdx)]\n [] PreserveFlags eval_Cpuid\n\n// The XGETBV instruction requires that OSXSAVE (in CPUID) is enabled.\n// We underspecify XGETBV here to only support fetching XCR0, which\n// is supported on any processor supporting the XGETBV instruction\nlet eval_Xgetbv (rcx:nat64) : option (nat64 & nat64) =\n if osxsave_enabled && rcx = 0 then Some (xgetbv rRax rcx, xgetbv rRdx rcx) else None\nval ins_Xgetbv :\n instr_dep\n [out (one64Reg rRax); out (one64Reg rRdx)]\n [one64Reg rRcx] PreserveFlags eval_Xgetbv\n\nlet check_avx (#a:Type0) (x:option a) : option a =\n if avx_enabled then x else None\n\nlet check_sse2 (#a:Type0) (x:option a) : option a =\n if sse2_enabled then x else None\n\nlet check_ssse3 (#a:Type0) (x:option a) : option a =\n if ssse3_enabled then x else None\n\nlet check_sse4_1 (#a:Type0) (x:option a) : option a =\n if sse4_1_enabled then x else None\n\nlet eval_Movdqu (src:quad32) : option quad32 = check_sse2 (Some src)\nval ins_Movdqu : instr_dep [out opXmm] [opXmm] PreserveFlags eval_Movdqu\n\nlet eval_Pxor (dst src:quad32) : option quad32 = check_sse2 (Some (quad32_xor dst src))\nval ins_Pxor : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pxor\n\nlet eval_VPxor (src1 src2:quad32) : option quad32 = check_avx (Some (quad32_xor src1 src2))\nval ins_VPxor : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPxor\n\nlet eval_Pand (dst src:quad32) : option quad32 = check_sse2 (Some (four_map2 (fun di si -> iand di si) dst src))\nval ins_Pand : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pand\n\nlet eval_Paddd_raw (src1 src2:quad32) : option quad32 =\n Some (Mkfour\n ((src1.lo0 + src2.lo0) % pow2_32)\n ((src1.lo1 + src2.lo1) % pow2_32)\n ((src1.hi2 + src2.hi2) % pow2_32)\n ((src1.hi3 + src2.hi3) % pow2_32))\n\nlet eval_Paddd (src1 src2:quad32) : option quad32 = check_sse2 (eval_Paddd_raw src1 src2)\nval ins_Paddd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Paddd\n\nlet eval_VPaddd (src1 src2:quad32) : option quad32 = check_avx (eval_Paddd_raw src1 src2)\nval ins_VPaddd : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPaddd\n\nlet eval_Pslld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishl i amt) dst) else None)\nval ins_Pslld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Pslld amt)\n\nlet eval_Psrld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishr i amt) dst) else None)\nval ins_Psrld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrld amt)\n\nlet eval_Psrldq (amt:int) (dst:quad32) : option quad32 = check_sse2 (\n if 0 <= amt && amt < 16 then\n let src_bytes = le_quad32_to_bytes dst in\n let zero_pad = Seq.create amt 0 in\n let remaining_bytes = slice src_bytes amt (length src_bytes) in\n Some (le_bytes_to_quad32 (append zero_pad remaining_bytes))\n else None)\nval ins_Psrldq (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrldq amt)\n\nlet eval_Palignr_raw (amount:nat8) (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if amount = 4 then\n Some (Mkfour src2.lo1 src2.hi2 src2.hi3 src1.lo0)\n else if amount = 8 then\n Some (Mkfour src2.hi2 src2.hi3 src1.lo0 src1.lo1)\n else None\n\nlet eval_Palignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_ssse3 (eval_Palignr_raw amount src1 src2)\nval ins_Palignr (amount:nat8) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Palignr amount)\n\nlet eval_VPalignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Palignr_raw amount src1 src2)\nval ins_VPalignr (amount:nat8) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPalignr amount)\n\nlet eval_Shufpd_raw (permutation:int) (src1 src2:quad32) : option quad32 =\n if 0 <= permutation && permutation < 4 then\n Some (Mkfour\n (if permutation % 2 = 0 then src1.lo0 else src1.hi2)\n (if permutation % 2 = 0 then src1.lo1 else src1.hi3)\n (if (permutation / 2) % 2 = 0 then src2.lo0 else src2.hi2)\n (if (permutation / 2) % 2 = 0 then src2.lo1 else src2.hi3))\n else None\n\nlet eval_Shufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_sse2 (eval_Shufpd_raw permutation src1 src2)\nval ins_Shufpd (permutation:int) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Shufpd permutation)\n\nlet eval_VShufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Shufpd_raw permutation src1 src2)\nval ins_VShufpd (permutation:int) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VShufpd permutation)\n\nlet is_full_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x04050607 &&\n q.hi3 = 0x00010203\n\nlet is_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x00010203 &&\n q.lo1 = 0x04050607 &&\n q.hi2 = 0x08090A0B &&\n q.hi3 = 0x0C0D0E0F\n\nlet is_high_dup_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet is_lower_upper_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x04050607 &&\n q.lo1 = 0x00010203 &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet eval_Pshufb_raw (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if is_full_byte_reversal_mask src2 then\n Some (reverse_bytes_quad32 src1)\n else if is_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3))\n else if is_high_dup_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else if is_lower_upper_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else None\n\nlet eval_Pshufb (src1 src2:quad32) : option quad32 = check_ssse3 (eval_Pshufb_raw src1 src2)\nval ins_Pshufb : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pshufb\n\nlet eval_VPshufb (src1 src2:quad32) : option quad32 = check_avx (eval_Pshufb_raw src1 src2)\nval ins_VPshufb : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPshufb\n\nlet eval_Pshufd (permutation:nat8) (src:quad32) : option quad32 = check_sse2 (\n let bits:bits_of_byte = byte_to_twobits permutation in\n Some (Mkfour\n (select_word src bits.lo0)\n (select_word src bits.lo1)\n (select_word src bits.hi2)\n (select_word src bits.hi3)))\nval ins_Pshufd (permutation:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_Pshufd permutation)\n\nlet eval_Pcmpeqd (dst src:quad32) : option quad32 = check_sse2 (\n let eq_result (b:bool):nat32 = if b then 0xFFFFFFFF else 0 in\n Some (Mkfour\n (eq_result (src.lo0 = dst.lo0))\n (eq_result (src.lo1 = dst.lo1))\n (eq_result (src.hi2 = dst.hi2))\n (eq_result (src.hi3 = dst.hi3))))\nval ins_Pcmpeqd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pcmpeqd\n\nlet eval_Pextrq (index:nat8) (src:quad32) : option nat64 = check_sse4_1 (\n let src_two = four_to_two_two src in\n Some (two_to_nat 32 (two_select src_two (index % 2))))\nval ins_Pextrq (index:nat8) : instr_dep [out op64] [opXmm] PreserveFlags (eval_Pextrq index)\n\nlet eval_Pinsrd (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat32 dst (src % pow2_32) (index % 4)))\nval ins_Pinsrd (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrd index)\n\nlet eval_Pinsrq (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat64_def dst src (index % 2)))\nval ins_Pinsrq (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrq index)\n\nlet eval_Pslldq_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the two very special cases we need\n if count = 4 then Some (Mkfour 0 src.lo0 src.lo1 src.hi2)\n else if count = 8 then Some (Mkfour 0 0 src.lo0 src.lo1)\n else None\nlet eval_VPslldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Pslldq_raw count src)\nval ins_VPslldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPslldq count)\n\nlet eval_Psrldq_8_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the one very special case we need\n if count = 8 then Some (Mkfour src.hi2 src.hi3 0 0)\n else None\nlet eval_VPsrldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Psrldq_8_raw count src)\nval ins_VPsrldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPsrldq count)\n\nlet eval_Pclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n let Mkfour a0 a1 a2 a3 = src1 in\n let Mkfour b0 b1 b2 b3 = src2 in\n let f x0 x1 y0 y1 =\n let x = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo x0 x1) in\n let y = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo y0 y1) in\n Vale.Math.Poly2.Bits_s.to_quad32 (Vale.Math.Poly2_s.mul x y)\n in\n if pclmulqdq_enabled then\n match imm with\n | 0 -> Some (f a0 a1 b0 b1)\n | 1 -> Some (f a2 a3 b0 b1)\n | 16 -> Some (f a0 a1 b2 b3)\n | 17 -> Some (f a2 a3 b2 b3)\n | _ -> None\n else None\nval ins_Pclmulqdq (imm:int) : instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Pclmulqdq imm)\n\nlet eval_VPclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Pclmulqdq imm src1 src2)\nval ins_VPclmulqdq (imm:int) : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPclmulqdq imm)\n\nlet eval_AESNI_enc (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.mix_columns_LE (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1))) src2)\n else None\nval ins_AESNI_enc : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc\n\nlet eval_VAESNI_enc (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc src1 src2)\nval ins_VAESNI_enc : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc\n\nlet eval_AESNI_enc_last (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1)) src2)\n else None\nval ins_AESNI_enc_last : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc_last\n\nlet eval_VAESNI_enc_last (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc_last src1 src2)\nval ins_VAESNI_enc_last : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc_last\n\nPartial implementation:\nlet eval_AESNI_dec (dst src: quad32) : option quad32 =\n \n"
},
{
"role": "assistant",
"content": "if aesni_enabled\nthen\n Some\n (quad32_xor (Vale.AES.AES_s.inv_mix_columns_LE (Vale.AES.AES_s.inv_sub_bytes (Vale.AES.AES_s.inv_shift_rows_LE\n dst)))\n src)\nelse None"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: FStar.Pervasives.Native.option Prims.unit\nEffect: Prims.Tot\n\nContext:\nmodule Vale.X64.Instructions_s\nopen FStar.Mul\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Two_s\nopen Vale.Def.Words.Four_s\nopen Vale.Def.Types_s\nopen Vale.X64.Machine_s\nopen Vale.X64.Instruction_s\nopen Vale.X64.CPU_Features_s\nopen Vale.X64.CryptoInstructions_s\nopen FStar.Seq.Base\n\nlet eval_Mov64 (src:nat64) : option nat64 = Some src\nval ins_Mov64 : instr_dep [out op64] [op64] PreserveFlags eval_Mov64\n\nlet eval_MovBe64 (src:nat64) : option nat64 = if movbe_enabled then Some (reverse_bytes_nat64 src) else None\nval ins_MovBe64 : instr_dep [out op64] [op64] PreserveFlags eval_MovBe64\n\nlet eval_Bswap64 (dst:nat64) : option nat64 = Some (reverse_bytes_nat64 dst)\nval ins_Bswap64 : instr_dep [inOut op64] [] PreserveFlags eval_Bswap64\n\nlet eval_Cmovc64 (dst src:nat64) (carry:bool) : option nat64 = Some (if carry then src else dst)\nval ins_Cmovc64 : instr_dep [inOut op64] [op64; opFlagsCf] PreserveFlags eval_Cmovc64\n\nlet eval_Add64 (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src in Some (sum >= pow2_64, sum % pow2_64)\nval ins_Add64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Add64\n\nlet eval_AddLea64 (src1 src2:nat64) : option nat64 = Some ((src1 + src2) % pow2_64)\nval ins_AddLea64 :instr_dep [out op64] [op64; op64] PreserveFlags eval_AddLea64\n\nlet eval_AddCarry64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_carry then 1 else 0) in\n Some (sum >= pow2_64, sum % pow2_64)\nval ins_AddCarry64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_AddCarry64\n\nlet eval_Adcx64_Adox64 (old_flag:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_flag then 1 else 0) in\n if adx_enabled then Some (sum >= pow2_64, sum % pow2_64) else None\nval ins_Adcx64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\nval ins_Adox64 : instr_dep [inOut opFlagsOf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\n\nlet eval_Sub64 (dst src:nat64) : option (bool & nat64) =\n let diff = dst - src in Some (diff < 0, diff % pow2_64)\nval ins_Sub64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Sub64\n\nlet eval_Sbb64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let diff = dst - (src + (if old_carry then 1 else 0)) in\n Some (diff < 0, diff % pow2_64) // We specify cf, but underspecify everything else (via HavocFlags)\nval ins_Sbb64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_Sbb64\n\nlet eval_Mul64 (rax src:nat64) : option (nat64 & nat64) =\n Some (FStar.UInt.mul_div #64 rax src, FStar.UInt.mul_mod #64 rax src)\nval ins_Mul64 : instr_dep [out (one64Reg rRdx); inOut (one64Reg rRax)] [op64] HavocFlags eval_Mul64\n\nlet eval_Mulx64 (rdx src:nat64) : option (nat64 & nat64) =\n let hi = FStar.UInt.mul_div #64 rdx src in\n let lo = FStar.UInt.mul_mod #64 rdx src in\n if bmi2_enabled then Some (hi, lo) else None\nval ins_Mulx64 : instr_dep [out op64; out op64] [one64Reg rRdx; op64] PreserveFlags eval_Mulx64\n\nlet eval_IMul64 (dst src:nat64) : option nat64 =\n Some (FStar.UInt.mul_mod #64 dst src)\nval ins_IMul64 : instr_dep [inOut op64] [op64] HavocFlags eval_IMul64\n\nlet eval_And64 (dst src:nat64) : option nat64 = Some (iand dst src)\nval ins_And64 : instr_dep [inOut op64] [op64] HavocFlags eval_And64\n\nlet eval_Xor64 (dst src:nat64) : option (nat64 & (bool & bool)) =\n Some (Vale.Def.Types_s.ixor dst src, (false, false))\nval ins_Xor64 : instr_dep [inOut op64; out opFlagsCf; out opFlagsOf] [op64] HavocFlags eval_Xor64\n\nlet eval_Shr64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishr dst amt) else None\nval ins_Shr64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shr64\n\nlet eval_Shl64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishl dst amt) else None\nval ins_Shl64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shl64\n\nlet eval_Cpuid (rax rcx:nat64) : option (nat64 & (nat64 & (nat64 & nat64))) =\n Some (cpuid rRax rax rcx, (cpuid rRbx rax rcx, (cpuid rRcx rax rcx, cpuid rRdx rax rcx)))\nval ins_Cpuid :\n instr_dep\n [inOut (one64Reg rRax); out (one64Reg rRbx); inOut (one64Reg rRcx); out (one64Reg rRdx)]\n [] PreserveFlags eval_Cpuid\n\n// The XGETBV instruction requires that OSXSAVE (in CPUID) is enabled.\n// We underspecify XGETBV here to only support fetching XCR0, which\n// is supported on any processor supporting the XGETBV instruction\nlet eval_Xgetbv (rcx:nat64) : option (nat64 & nat64) =\n if osxsave_enabled && rcx = 0 then Some (xgetbv rRax rcx, xgetbv rRdx rcx) else None\nval ins_Xgetbv :\n instr_dep\n [out (one64Reg rRax); out (one64Reg rRdx)]\n [one64Reg rRcx] PreserveFlags eval_Xgetbv\n\nlet check_avx (#a:Type0) (x:option a) : option a =\n if avx_enabled then x else None\n\nlet check_sse2 (#a:Type0) (x:option a) : option a =\n if sse2_enabled then x else None\n\nlet check_ssse3 (#a:Type0) (x:option a) : option a =\n if ssse3_enabled then x else None\n\nlet check_sse4_1 (#a:Type0) (x:option a) : option a =\n if sse4_1_enabled then x else None\n\nlet eval_Movdqu (src:quad32) : option quad32 = check_sse2 (Some src)\nval ins_Movdqu : instr_dep [out opXmm] [opXmm] PreserveFlags eval_Movdqu\n\nlet eval_Pxor (dst src:quad32) : option quad32 = check_sse2 (Some (quad32_xor dst src))\nval ins_Pxor : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pxor\n\nlet eval_VPxor (src1 src2:quad32) : option quad32 = check_avx (Some (quad32_xor src1 src2))\nval ins_VPxor : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPxor\n\nlet eval_Pand (dst src:quad32) : option quad32 = check_sse2 (Some (four_map2 (fun di si -> iand di si) dst src))\nval ins_Pand : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pand\n\nlet eval_Paddd_raw (src1 src2:quad32) : option quad32 =\n Some (Mkfour\n ((src1.lo0 + src2.lo0) % pow2_32)\n ((src1.lo1 + src2.lo1) % pow2_32)\n ((src1.hi2 + src2.hi2) % pow2_32)\n ((src1.hi3 + src2.hi3) % pow2_32))\n\nlet eval_Paddd (src1 src2:quad32) : option quad32 = check_sse2 (eval_Paddd_raw src1 src2)\nval ins_Paddd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Paddd\n\nlet eval_VPaddd (src1 src2:quad32) : option quad32 = check_avx (eval_Paddd_raw src1 src2)\nval ins_VPaddd : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPaddd\n\nlet eval_Pslld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishl i amt) dst) else None)\nval ins_Pslld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Pslld amt)\n\nlet eval_Psrld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishr i amt) dst) else None)\nval ins_Psrld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrld amt)\n\nlet eval_Psrldq (amt:int) (dst:quad32) : option quad32 = check_sse2 (\n if 0 <= amt && amt < 16 then\n let src_bytes = le_quad32_to_bytes dst in\n let zero_pad = Seq.create amt 0 in\n let remaining_bytes = slice src_bytes amt (length src_bytes) in\n Some (le_bytes_to_quad32 (append zero_pad remaining_bytes))\n else None)\nval ins_Psrldq (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrldq amt)\n\nlet eval_Palignr_raw (amount:nat8) (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if amount = 4 then\n Some (Mkfour src2.lo1 src2.hi2 src2.hi3 src1.lo0)\n else if amount = 8 then\n Some (Mkfour src2.hi2 src2.hi3 src1.lo0 src1.lo1)\n else None\n\nlet eval_Palignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_ssse3 (eval_Palignr_raw amount src1 src2)\nval ins_Palignr (amount:nat8) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Palignr amount)\n\nlet eval_VPalignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Palignr_raw amount src1 src2)\nval ins_VPalignr (amount:nat8) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPalignr amount)\n\nlet eval_Shufpd_raw (permutation:int) (src1 src2:quad32) : option quad32 =\n if 0 <= permutation && permutation < 4 then\n Some (Mkfour\n (if permutation % 2 = 0 then src1.lo0 else src1.hi2)\n (if permutation % 2 = 0 then src1.lo1 else src1.hi3)\n (if (permutation / 2) % 2 = 0 then src2.lo0 else src2.hi2)\n (if (permutation / 2) % 2 = 0 then src2.lo1 else src2.hi3))\n else None\n\nlet eval_Shufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_sse2 (eval_Shufpd_raw permutation src1 src2)\nval ins_Shufpd (permutation:int) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Shufpd permutation)\n\nlet eval_VShufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Shufpd_raw permutation src1 src2)\nval ins_VShufpd (permutation:int) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VShufpd permutation)\n\nlet is_full_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x04050607 &&\n q.hi3 = 0x00010203\n\nlet is_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x00010203 &&\n q.lo1 = 0x04050607 &&\n q.hi2 = 0x08090A0B &&\n q.hi3 = 0x0C0D0E0F\n\nlet is_high_dup_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet is_lower_upper_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x04050607 &&\n q.lo1 = 0x00010203 &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet eval_Pshufb_raw (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if is_full_byte_reversal_mask src2 then\n Some (reverse_bytes_quad32 src1)\n else if is_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3))\n else if is_high_dup_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else if is_lower_upper_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else None\n\nlet eval_Pshufb (src1 src2:quad32) : option quad32 = check_ssse3 (eval_Pshufb_raw src1 src2)\nval ins_Pshufb : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pshufb\n\nlet eval_VPshufb (src1 src2:quad32) : option quad32 = check_avx (eval_Pshufb_raw src1 src2)\nval ins_VPshufb : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPshufb\n\nlet eval_Pshufd (permutation:nat8) (src:quad32) : option quad32 = check_sse2 (\n let bits:bits_of_byte = byte_to_twobits permutation in\n Some (Mkfour\n (select_word src bits.lo0)\n (select_word src bits.lo1)\n (select_word src bits.hi2)\n (select_word src bits.hi3)))\nval ins_Pshufd (permutation:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_Pshufd permutation)\n\nlet eval_Pcmpeqd (dst src:quad32) : option quad32 = check_sse2 (\n let eq_result (b:bool):nat32 = if b then 0xFFFFFFFF else 0 in\n Some (Mkfour\n (eq_result (src.lo0 = dst.lo0))\n (eq_result (src.lo1 = dst.lo1))\n (eq_result (src.hi2 = dst.hi2))\n (eq_result (src.hi3 = dst.hi3))))\nval ins_Pcmpeqd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pcmpeqd\n\nlet eval_Pextrq (index:nat8) (src:quad32) : option nat64 = check_sse4_1 (\n let src_two = four_to_two_two src in\n Some (two_to_nat 32 (two_select src_two (index % 2))))\nval ins_Pextrq (index:nat8) : instr_dep [out op64] [opXmm] PreserveFlags (eval_Pextrq index)\n\nlet eval_Pinsrd (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat32 dst (src % pow2_32) (index % 4)))\nval ins_Pinsrd (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrd index)\n\nlet eval_Pinsrq (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat64_def dst src (index % 2)))\nval ins_Pinsrq (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrq index)\n\nlet eval_Pslldq_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the two very special cases we need\n if count = 4 then Some (Mkfour 0 src.lo0 src.lo1 src.hi2)\n else if count = 8 then Some (Mkfour 0 0 src.lo0 src.lo1)\n else None\nlet eval_VPslldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Pslldq_raw count src)\nval ins_VPslldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPslldq count)\n\nlet eval_Psrldq_8_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the one very special case we need\n if count = 8 then Some (Mkfour src.hi2 src.hi3 0 0)\n else None\nlet eval_VPsrldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Psrldq_8_raw count src)\nval ins_VPsrldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPsrldq count)\n\nlet eval_Pclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n let Mkfour a0 a1 a2 a3 = src1 in\n let Mkfour b0 b1 b2 b3 = src2 in\n let f x0 x1 y0 y1 =\n let x = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo x0 x1) in\n let y = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo y0 y1) in\n Vale.Math.Poly2.Bits_s.to_quad32 (Vale.Math.Poly2_s.mul x y)\n in\n if pclmulqdq_enabled then\n match imm with\n | 0 -> Some (f a0 a1 b0 b1)\n | 1 -> Some (f a2 a3 b0 b1)\n | 16 -> Some (f a0 a1 b2 b3)\n | 17 -> Some (f a2 a3 b2 b3)\n | _ -> None\n else None\nval ins_Pclmulqdq (imm:int) : instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Pclmulqdq imm)\n\nlet eval_VPclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Pclmulqdq imm src1 src2)\nval ins_VPclmulqdq (imm:int) : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPclmulqdq imm)\n\nlet eval_AESNI_enc (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.mix_columns_LE (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1))) src2)\n else None\nval ins_AESNI_enc : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc\n\nlet eval_VAESNI_enc (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc src1 src2)\nval ins_VAESNI_enc : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc\n\nlet eval_AESNI_enc_last (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1)) src2)\n else None\nval ins_AESNI_enc_last : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc_last\n\nlet eval_VAESNI_enc_last (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc_last src1 src2)\nval ins_VAESNI_enc_last : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc_last\n\nlet eval_AESNI_dec (dst src:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.inv_mix_columns_LE (Vale.AES.AES_s.inv_sub_bytes (Vale.AES.AES_s.inv_shift_rows_LE dst))) src)\n else None\nval ins_AESNI_dec : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_dec\n\nlet eval_AESNI_dec_last (dst src:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.inv_sub_bytes (Vale.AES.AES_s.inv_shift_rows_LE dst)) src)\n else None\nval ins_AESNI_dec_last : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_dec_last\n\nlet eval_AESNI_imc (src:quad32) : option quad32 =\n if aesni_enabled then Some (Vale.AES.AES_s.inv_mix_columns_LE src) else None\nval ins_AESNI_imc : instr_dep [out opXmm] [opXmm] PreserveFlags eval_AESNI_imc\n\nlet eval_AESNI_keygen_assist (imm:nat8) (src:quad32) : option quad32 =\n if aesni_enabled then\n Some (Mkfour\n (Vale.AES.AES_s.sub_word src.lo1)\n (ixor (Vale.AES.AES_s.rot_word_LE (Vale.AES.AES_s.sub_word src.lo1)) imm)\n (Vale.AES.AES_s.sub_word src.hi3)\n (ixor (Vale.AES.AES_s.rot_word_LE (Vale.AES.AES_s.sub_word src.hi3)) imm))\n else None\nval ins_AESNI_keygen_assist (imm:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_AESNI_keygen_assist imm)\n\nlet eval_SHA256_rnds2 (src1 src2 wk:quad32) : option quad32 =\n if sha_enabled then Some (sha256_rnds2_spec src1 src2 wk) else None\nval ins_SHA256_rnds2 :\n instr_dep [inOut opXmm] [opXmm; oneXmm (OReg 0)] PreserveFlags eval_SHA256_rnds2\n\nlet eval_SHA256_msg1 (src1 src2:quad32) : option quad32 =\n if sha_enabled then Some (sha256_msg1_spec src1 src2) else None\nval ins_SHA256_msg1 : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_SHA256_msg1\n\nlet eval_SHA256_msg2 (src1 src2:quad32) : option quad32 =\n if sha_enabled then Some (sha256_msg2_spec src1 src2) else None\nval ins_SHA256_msg2 : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_SHA256_msg2\n\nlet eval_Ghost : option unit = Some ()\nval ins_Ghost : instr_dep [] [] PreserveFlags eval_Ghost\n\nlet eval_Comment : option unit = Some ()\nval ins_Comment (_:string) : instr_dep [] [] PreserveFlags eval_Comment\n\nPartial implementation:\nlet eval_LargeComment:option unit =\n \n"
},
{
"role": "assistant",
"content": "Some ()"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: src1: Vale.X64.Machine_s.quad32 -> src2: Vale.X64.Machine_s.quad32 -> wk: Vale.X64.Machine_s.quad32\n -> FStar.Pervasives.Native.option Vale.X64.Machine_s.quad32\nEffect: Prims.Tot\n\nContext:\nmodule Vale.X64.Instructions_s\nopen FStar.Mul\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Two_s\nopen Vale.Def.Words.Four_s\nopen Vale.Def.Types_s\nopen Vale.X64.Machine_s\nopen Vale.X64.Instruction_s\nopen Vale.X64.CPU_Features_s\nopen Vale.X64.CryptoInstructions_s\nopen FStar.Seq.Base\n\nlet eval_Mov64 (src:nat64) : option nat64 = Some src\nval ins_Mov64 : instr_dep [out op64] [op64] PreserveFlags eval_Mov64\n\nlet eval_MovBe64 (src:nat64) : option nat64 = if movbe_enabled then Some (reverse_bytes_nat64 src) else None\nval ins_MovBe64 : instr_dep [out op64] [op64] PreserveFlags eval_MovBe64\n\nlet eval_Bswap64 (dst:nat64) : option nat64 = Some (reverse_bytes_nat64 dst)\nval ins_Bswap64 : instr_dep [inOut op64] [] PreserveFlags eval_Bswap64\n\nlet eval_Cmovc64 (dst src:nat64) (carry:bool) : option nat64 = Some (if carry then src else dst)\nval ins_Cmovc64 : instr_dep [inOut op64] [op64; opFlagsCf] PreserveFlags eval_Cmovc64\n\nlet eval_Add64 (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src in Some (sum >= pow2_64, sum % pow2_64)\nval ins_Add64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Add64\n\nlet eval_AddLea64 (src1 src2:nat64) : option nat64 = Some ((src1 + src2) % pow2_64)\nval ins_AddLea64 :instr_dep [out op64] [op64; op64] PreserveFlags eval_AddLea64\n\nlet eval_AddCarry64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_carry then 1 else 0) in\n Some (sum >= pow2_64, sum % pow2_64)\nval ins_AddCarry64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_AddCarry64\n\nlet eval_Adcx64_Adox64 (old_flag:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_flag then 1 else 0) in\n if adx_enabled then Some (sum >= pow2_64, sum % pow2_64) else None\nval ins_Adcx64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\nval ins_Adox64 : instr_dep [inOut opFlagsOf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\n\nlet eval_Sub64 (dst src:nat64) : option (bool & nat64) =\n let diff = dst - src in Some (diff < 0, diff % pow2_64)\nval ins_Sub64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Sub64\n\nlet eval_Sbb64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let diff = dst - (src + (if old_carry then 1 else 0)) in\n Some (diff < 0, diff % pow2_64) // We specify cf, but underspecify everything else (via HavocFlags)\nval ins_Sbb64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_Sbb64\n\nlet eval_Mul64 (rax src:nat64) : option (nat64 & nat64) =\n Some (FStar.UInt.mul_div #64 rax src, FStar.UInt.mul_mod #64 rax src)\nval ins_Mul64 : instr_dep [out (one64Reg rRdx); inOut (one64Reg rRax)] [op64] HavocFlags eval_Mul64\n\nlet eval_Mulx64 (rdx src:nat64) : option (nat64 & nat64) =\n let hi = FStar.UInt.mul_div #64 rdx src in\n let lo = FStar.UInt.mul_mod #64 rdx src in\n if bmi2_enabled then Some (hi, lo) else None\nval ins_Mulx64 : instr_dep [out op64; out op64] [one64Reg rRdx; op64] PreserveFlags eval_Mulx64\n\nlet eval_IMul64 (dst src:nat64) : option nat64 =\n Some (FStar.UInt.mul_mod #64 dst src)\nval ins_IMul64 : instr_dep [inOut op64] [op64] HavocFlags eval_IMul64\n\nlet eval_And64 (dst src:nat64) : option nat64 = Some (iand dst src)\nval ins_And64 : instr_dep [inOut op64] [op64] HavocFlags eval_And64\n\nlet eval_Xor64 (dst src:nat64) : option (nat64 & (bool & bool)) =\n Some (Vale.Def.Types_s.ixor dst src, (false, false))\nval ins_Xor64 : instr_dep [inOut op64; out opFlagsCf; out opFlagsOf] [op64] HavocFlags eval_Xor64\n\nlet eval_Shr64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishr dst amt) else None\nval ins_Shr64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shr64\n\nlet eval_Shl64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishl dst amt) else None\nval ins_Shl64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shl64\n\nlet eval_Cpuid (rax rcx:nat64) : option (nat64 & (nat64 & (nat64 & nat64))) =\n Some (cpuid rRax rax rcx, (cpuid rRbx rax rcx, (cpuid rRcx rax rcx, cpuid rRdx rax rcx)))\nval ins_Cpuid :\n instr_dep\n [inOut (one64Reg rRax); out (one64Reg rRbx); inOut (one64Reg rRcx); out (one64Reg rRdx)]\n [] PreserveFlags eval_Cpuid\n\n// The XGETBV instruction requires that OSXSAVE (in CPUID) is enabled.\n// We underspecify XGETBV here to only support fetching XCR0, which\n// is supported on any processor supporting the XGETBV instruction\nlet eval_Xgetbv (rcx:nat64) : option (nat64 & nat64) =\n if osxsave_enabled && rcx = 0 then Some (xgetbv rRax rcx, xgetbv rRdx rcx) else None\nval ins_Xgetbv :\n instr_dep\n [out (one64Reg rRax); out (one64Reg rRdx)]\n [one64Reg rRcx] PreserveFlags eval_Xgetbv\n\nlet check_avx (#a:Type0) (x:option a) : option a =\n if avx_enabled then x else None\n\nlet check_sse2 (#a:Type0) (x:option a) : option a =\n if sse2_enabled then x else None\n\nlet check_ssse3 (#a:Type0) (x:option a) : option a =\n if ssse3_enabled then x else None\n\nlet check_sse4_1 (#a:Type0) (x:option a) : option a =\n if sse4_1_enabled then x else None\n\nlet eval_Movdqu (src:quad32) : option quad32 = check_sse2 (Some src)\nval ins_Movdqu : instr_dep [out opXmm] [opXmm] PreserveFlags eval_Movdqu\n\nlet eval_Pxor (dst src:quad32) : option quad32 = check_sse2 (Some (quad32_xor dst src))\nval ins_Pxor : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pxor\n\nlet eval_VPxor (src1 src2:quad32) : option quad32 = check_avx (Some (quad32_xor src1 src2))\nval ins_VPxor : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPxor\n\nlet eval_Pand (dst src:quad32) : option quad32 = check_sse2 (Some (four_map2 (fun di si -> iand di si) dst src))\nval ins_Pand : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pand\n\nlet eval_Paddd_raw (src1 src2:quad32) : option quad32 =\n Some (Mkfour\n ((src1.lo0 + src2.lo0) % pow2_32)\n ((src1.lo1 + src2.lo1) % pow2_32)\n ((src1.hi2 + src2.hi2) % pow2_32)\n ((src1.hi3 + src2.hi3) % pow2_32))\n\nlet eval_Paddd (src1 src2:quad32) : option quad32 = check_sse2 (eval_Paddd_raw src1 src2)\nval ins_Paddd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Paddd\n\nlet eval_VPaddd (src1 src2:quad32) : option quad32 = check_avx (eval_Paddd_raw src1 src2)\nval ins_VPaddd : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPaddd\n\nlet eval_Pslld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishl i amt) dst) else None)\nval ins_Pslld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Pslld amt)\n\nlet eval_Psrld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishr i amt) dst) else None)\nval ins_Psrld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrld amt)\n\nlet eval_Psrldq (amt:int) (dst:quad32) : option quad32 = check_sse2 (\n if 0 <= amt && amt < 16 then\n let src_bytes = le_quad32_to_bytes dst in\n let zero_pad = Seq.create amt 0 in\n let remaining_bytes = slice src_bytes amt (length src_bytes) in\n Some (le_bytes_to_quad32 (append zero_pad remaining_bytes))\n else None)\nval ins_Psrldq (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrldq amt)\n\nlet eval_Palignr_raw (amount:nat8) (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if amount = 4 then\n Some (Mkfour src2.lo1 src2.hi2 src2.hi3 src1.lo0)\n else if amount = 8 then\n Some (Mkfour src2.hi2 src2.hi3 src1.lo0 src1.lo1)\n else None\n\nlet eval_Palignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_ssse3 (eval_Palignr_raw amount src1 src2)\nval ins_Palignr (amount:nat8) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Palignr amount)\n\nlet eval_VPalignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Palignr_raw amount src1 src2)\nval ins_VPalignr (amount:nat8) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPalignr amount)\n\nlet eval_Shufpd_raw (permutation:int) (src1 src2:quad32) : option quad32 =\n if 0 <= permutation && permutation < 4 then\n Some (Mkfour\n (if permutation % 2 = 0 then src1.lo0 else src1.hi2)\n (if permutation % 2 = 0 then src1.lo1 else src1.hi3)\n (if (permutation / 2) % 2 = 0 then src2.lo0 else src2.hi2)\n (if (permutation / 2) % 2 = 0 then src2.lo1 else src2.hi3))\n else None\n\nlet eval_Shufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_sse2 (eval_Shufpd_raw permutation src1 src2)\nval ins_Shufpd (permutation:int) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Shufpd permutation)\n\nlet eval_VShufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Shufpd_raw permutation src1 src2)\nval ins_VShufpd (permutation:int) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VShufpd permutation)\n\nlet is_full_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x04050607 &&\n q.hi3 = 0x00010203\n\nlet is_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x00010203 &&\n q.lo1 = 0x04050607 &&\n q.hi2 = 0x08090A0B &&\n q.hi3 = 0x0C0D0E0F\n\nlet is_high_dup_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet is_lower_upper_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x04050607 &&\n q.lo1 = 0x00010203 &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet eval_Pshufb_raw (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if is_full_byte_reversal_mask src2 then\n Some (reverse_bytes_quad32 src1)\n else if is_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3))\n else if is_high_dup_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else if is_lower_upper_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else None\n\nlet eval_Pshufb (src1 src2:quad32) : option quad32 = check_ssse3 (eval_Pshufb_raw src1 src2)\nval ins_Pshufb : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pshufb\n\nlet eval_VPshufb (src1 src2:quad32) : option quad32 = check_avx (eval_Pshufb_raw src1 src2)\nval ins_VPshufb : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPshufb\n\nlet eval_Pshufd (permutation:nat8) (src:quad32) : option quad32 = check_sse2 (\n let bits:bits_of_byte = byte_to_twobits permutation in\n Some (Mkfour\n (select_word src bits.lo0)\n (select_word src bits.lo1)\n (select_word src bits.hi2)\n (select_word src bits.hi3)))\nval ins_Pshufd (permutation:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_Pshufd permutation)\n\nlet eval_Pcmpeqd (dst src:quad32) : option quad32 = check_sse2 (\n let eq_result (b:bool):nat32 = if b then 0xFFFFFFFF else 0 in\n Some (Mkfour\n (eq_result (src.lo0 = dst.lo0))\n (eq_result (src.lo1 = dst.lo1))\n (eq_result (src.hi2 = dst.hi2))\n (eq_result (src.hi3 = dst.hi3))))\nval ins_Pcmpeqd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pcmpeqd\n\nlet eval_Pextrq (index:nat8) (src:quad32) : option nat64 = check_sse4_1 (\n let src_two = four_to_two_two src in\n Some (two_to_nat 32 (two_select src_two (index % 2))))\nval ins_Pextrq (index:nat8) : instr_dep [out op64] [opXmm] PreserveFlags (eval_Pextrq index)\n\nlet eval_Pinsrd (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat32 dst (src % pow2_32) (index % 4)))\nval ins_Pinsrd (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrd index)\n\nlet eval_Pinsrq (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat64_def dst src (index % 2)))\nval ins_Pinsrq (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrq index)\n\nlet eval_Pslldq_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the two very special cases we need\n if count = 4 then Some (Mkfour 0 src.lo0 src.lo1 src.hi2)\n else if count = 8 then Some (Mkfour 0 0 src.lo0 src.lo1)\n else None\nlet eval_VPslldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Pslldq_raw count src)\nval ins_VPslldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPslldq count)\n\nlet eval_Psrldq_8_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the one very special case we need\n if count = 8 then Some (Mkfour src.hi2 src.hi3 0 0)\n else None\nlet eval_VPsrldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Psrldq_8_raw count src)\nval ins_VPsrldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPsrldq count)\n\nlet eval_Pclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n let Mkfour a0 a1 a2 a3 = src1 in\n let Mkfour b0 b1 b2 b3 = src2 in\n let f x0 x1 y0 y1 =\n let x = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo x0 x1) in\n let y = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo y0 y1) in\n Vale.Math.Poly2.Bits_s.to_quad32 (Vale.Math.Poly2_s.mul x y)\n in\n if pclmulqdq_enabled then\n match imm with\n | 0 -> Some (f a0 a1 b0 b1)\n | 1 -> Some (f a2 a3 b0 b1)\n | 16 -> Some (f a0 a1 b2 b3)\n | 17 -> Some (f a2 a3 b2 b3)\n | _ -> None\n else None\nval ins_Pclmulqdq (imm:int) : instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Pclmulqdq imm)\n\nlet eval_VPclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Pclmulqdq imm src1 src2)\nval ins_VPclmulqdq (imm:int) : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPclmulqdq imm)\n\nlet eval_AESNI_enc (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.mix_columns_LE (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1))) src2)\n else None\nval ins_AESNI_enc : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc\n\nlet eval_VAESNI_enc (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc src1 src2)\nval ins_VAESNI_enc : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc\n\nlet eval_AESNI_enc_last (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1)) src2)\n else None\nval ins_AESNI_enc_last : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc_last\n\nlet eval_VAESNI_enc_last (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc_last src1 src2)\nval ins_VAESNI_enc_last : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc_last\n\nlet eval_AESNI_dec (dst src:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.inv_mix_columns_LE (Vale.AES.AES_s.inv_sub_bytes (Vale.AES.AES_s.inv_shift_rows_LE dst))) src)\n else None\nval ins_AESNI_dec : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_dec\n\nlet eval_AESNI_dec_last (dst src:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.inv_sub_bytes (Vale.AES.AES_s.inv_shift_rows_LE dst)) src)\n else None\nval ins_AESNI_dec_last : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_dec_last\n\nlet eval_AESNI_imc (src:quad32) : option quad32 =\n if aesni_enabled then Some (Vale.AES.AES_s.inv_mix_columns_LE src) else None\nval ins_AESNI_imc : instr_dep [out opXmm] [opXmm] PreserveFlags eval_AESNI_imc\n\nlet eval_AESNI_keygen_assist (imm:nat8) (src:quad32) : option quad32 =\n if aesni_enabled then\n Some (Mkfour\n (Vale.AES.AES_s.sub_word src.lo1)\n (ixor (Vale.AES.AES_s.rot_word_LE (Vale.AES.AES_s.sub_word src.lo1)) imm)\n (Vale.AES.AES_s.sub_word src.hi3)\n (ixor (Vale.AES.AES_s.rot_word_LE (Vale.AES.AES_s.sub_word src.hi3)) imm))\n else None\nval ins_AESNI_keygen_assist (imm:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_AESNI_keygen_assist imm)\n\nPartial implementation:\nlet eval_SHA256_rnds2 (src1 src2 wk: quad32) : option quad32 =\n \n"
},
{
"role": "assistant",
"content": "if sha_enabled then Some (sha256_rnds2_spec src1 src2 wk) else None"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: FStar.Pervasives.Native.option Prims.unit\nEffect: Prims.Tot\n\nContext:\nmodule Vale.X64.Instructions_s\nopen FStar.Mul\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Two_s\nopen Vale.Def.Words.Four_s\nopen Vale.Def.Types_s\nopen Vale.X64.Machine_s\nopen Vale.X64.Instruction_s\nopen Vale.X64.CPU_Features_s\nopen Vale.X64.CryptoInstructions_s\nopen FStar.Seq.Base\n\nlet eval_Mov64 (src:nat64) : option nat64 = Some src\nval ins_Mov64 : instr_dep [out op64] [op64] PreserveFlags eval_Mov64\n\nlet eval_MovBe64 (src:nat64) : option nat64 = if movbe_enabled then Some (reverse_bytes_nat64 src) else None\nval ins_MovBe64 : instr_dep [out op64] [op64] PreserveFlags eval_MovBe64\n\nlet eval_Bswap64 (dst:nat64) : option nat64 = Some (reverse_bytes_nat64 dst)\nval ins_Bswap64 : instr_dep [inOut op64] [] PreserveFlags eval_Bswap64\n\nlet eval_Cmovc64 (dst src:nat64) (carry:bool) : option nat64 = Some (if carry then src else dst)\nval ins_Cmovc64 : instr_dep [inOut op64] [op64; opFlagsCf] PreserveFlags eval_Cmovc64\n\nlet eval_Add64 (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src in Some (sum >= pow2_64, sum % pow2_64)\nval ins_Add64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Add64\n\nlet eval_AddLea64 (src1 src2:nat64) : option nat64 = Some ((src1 + src2) % pow2_64)\nval ins_AddLea64 :instr_dep [out op64] [op64; op64] PreserveFlags eval_AddLea64\n\nlet eval_AddCarry64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_carry then 1 else 0) in\n Some (sum >= pow2_64, sum % pow2_64)\nval ins_AddCarry64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_AddCarry64\n\nlet eval_Adcx64_Adox64 (old_flag:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_flag then 1 else 0) in\n if adx_enabled then Some (sum >= pow2_64, sum % pow2_64) else None\nval ins_Adcx64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\nval ins_Adox64 : instr_dep [inOut opFlagsOf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\n\nlet eval_Sub64 (dst src:nat64) : option (bool & nat64) =\n let diff = dst - src in Some (diff < 0, diff % pow2_64)\nval ins_Sub64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Sub64\n\nlet eval_Sbb64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let diff = dst - (src + (if old_carry then 1 else 0)) in\n Some (diff < 0, diff % pow2_64) // We specify cf, but underspecify everything else (via HavocFlags)\nval ins_Sbb64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_Sbb64\n\nlet eval_Mul64 (rax src:nat64) : option (nat64 & nat64) =\n Some (FStar.UInt.mul_div #64 rax src, FStar.UInt.mul_mod #64 rax src)\nval ins_Mul64 : instr_dep [out (one64Reg rRdx); inOut (one64Reg rRax)] [op64] HavocFlags eval_Mul64\n\nlet eval_Mulx64 (rdx src:nat64) : option (nat64 & nat64) =\n let hi = FStar.UInt.mul_div #64 rdx src in\n let lo = FStar.UInt.mul_mod #64 rdx src in\n if bmi2_enabled then Some (hi, lo) else None\nval ins_Mulx64 : instr_dep [out op64; out op64] [one64Reg rRdx; op64] PreserveFlags eval_Mulx64\n\nlet eval_IMul64 (dst src:nat64) : option nat64 =\n Some (FStar.UInt.mul_mod #64 dst src)\nval ins_IMul64 : instr_dep [inOut op64] [op64] HavocFlags eval_IMul64\n\nlet eval_And64 (dst src:nat64) : option nat64 = Some (iand dst src)\nval ins_And64 : instr_dep [inOut op64] [op64] HavocFlags eval_And64\n\nlet eval_Xor64 (dst src:nat64) : option (nat64 & (bool & bool)) =\n Some (Vale.Def.Types_s.ixor dst src, (false, false))\nval ins_Xor64 : instr_dep [inOut op64; out opFlagsCf; out opFlagsOf] [op64] HavocFlags eval_Xor64\n\nlet eval_Shr64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishr dst amt) else None\nval ins_Shr64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shr64\n\nlet eval_Shl64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishl dst amt) else None\nval ins_Shl64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shl64\n\nlet eval_Cpuid (rax rcx:nat64) : option (nat64 & (nat64 & (nat64 & nat64))) =\n Some (cpuid rRax rax rcx, (cpuid rRbx rax rcx, (cpuid rRcx rax rcx, cpuid rRdx rax rcx)))\nval ins_Cpuid :\n instr_dep\n [inOut (one64Reg rRax); out (one64Reg rRbx); inOut (one64Reg rRcx); out (one64Reg rRdx)]\n [] PreserveFlags eval_Cpuid\n\n// The XGETBV instruction requires that OSXSAVE (in CPUID) is enabled.\n// We underspecify XGETBV here to only support fetching XCR0, which\n// is supported on any processor supporting the XGETBV instruction\nlet eval_Xgetbv (rcx:nat64) : option (nat64 & nat64) =\n if osxsave_enabled && rcx = 0 then Some (xgetbv rRax rcx, xgetbv rRdx rcx) else None\nval ins_Xgetbv :\n instr_dep\n [out (one64Reg rRax); out (one64Reg rRdx)]\n [one64Reg rRcx] PreserveFlags eval_Xgetbv\n\nlet check_avx (#a:Type0) (x:option a) : option a =\n if avx_enabled then x else None\n\nlet check_sse2 (#a:Type0) (x:option a) : option a =\n if sse2_enabled then x else None\n\nlet check_ssse3 (#a:Type0) (x:option a) : option a =\n if ssse3_enabled then x else None\n\nlet check_sse4_1 (#a:Type0) (x:option a) : option a =\n if sse4_1_enabled then x else None\n\nlet eval_Movdqu (src:quad32) : option quad32 = check_sse2 (Some src)\nval ins_Movdqu : instr_dep [out opXmm] [opXmm] PreserveFlags eval_Movdqu\n\nlet eval_Pxor (dst src:quad32) : option quad32 = check_sse2 (Some (quad32_xor dst src))\nval ins_Pxor : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pxor\n\nlet eval_VPxor (src1 src2:quad32) : option quad32 = check_avx (Some (quad32_xor src1 src2))\nval ins_VPxor : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPxor\n\nlet eval_Pand (dst src:quad32) : option quad32 = check_sse2 (Some (four_map2 (fun di si -> iand di si) dst src))\nval ins_Pand : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pand\n\nlet eval_Paddd_raw (src1 src2:quad32) : option quad32 =\n Some (Mkfour\n ((src1.lo0 + src2.lo0) % pow2_32)\n ((src1.lo1 + src2.lo1) % pow2_32)\n ((src1.hi2 + src2.hi2) % pow2_32)\n ((src1.hi3 + src2.hi3) % pow2_32))\n\nlet eval_Paddd (src1 src2:quad32) : option quad32 = check_sse2 (eval_Paddd_raw src1 src2)\nval ins_Paddd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Paddd\n\nlet eval_VPaddd (src1 src2:quad32) : option quad32 = check_avx (eval_Paddd_raw src1 src2)\nval ins_VPaddd : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPaddd\n\nlet eval_Pslld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishl i amt) dst) else None)\nval ins_Pslld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Pslld amt)\n\nlet eval_Psrld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishr i amt) dst) else None)\nval ins_Psrld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrld amt)\n\nlet eval_Psrldq (amt:int) (dst:quad32) : option quad32 = check_sse2 (\n if 0 <= amt && amt < 16 then\n let src_bytes = le_quad32_to_bytes dst in\n let zero_pad = Seq.create amt 0 in\n let remaining_bytes = slice src_bytes amt (length src_bytes) in\n Some (le_bytes_to_quad32 (append zero_pad remaining_bytes))\n else None)\nval ins_Psrldq (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrldq amt)\n\nlet eval_Palignr_raw (amount:nat8) (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if amount = 4 then\n Some (Mkfour src2.lo1 src2.hi2 src2.hi3 src1.lo0)\n else if amount = 8 then\n Some (Mkfour src2.hi2 src2.hi3 src1.lo0 src1.lo1)\n else None\n\nlet eval_Palignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_ssse3 (eval_Palignr_raw amount src1 src2)\nval ins_Palignr (amount:nat8) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Palignr amount)\n\nlet eval_VPalignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Palignr_raw amount src1 src2)\nval ins_VPalignr (amount:nat8) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPalignr amount)\n\nlet eval_Shufpd_raw (permutation:int) (src1 src2:quad32) : option quad32 =\n if 0 <= permutation && permutation < 4 then\n Some (Mkfour\n (if permutation % 2 = 0 then src1.lo0 else src1.hi2)\n (if permutation % 2 = 0 then src1.lo1 else src1.hi3)\n (if (permutation / 2) % 2 = 0 then src2.lo0 else src2.hi2)\n (if (permutation / 2) % 2 = 0 then src2.lo1 else src2.hi3))\n else None\n\nlet eval_Shufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_sse2 (eval_Shufpd_raw permutation src1 src2)\nval ins_Shufpd (permutation:int) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Shufpd permutation)\n\nlet eval_VShufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Shufpd_raw permutation src1 src2)\nval ins_VShufpd (permutation:int) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VShufpd permutation)\n\nlet is_full_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x04050607 &&\n q.hi3 = 0x00010203\n\nlet is_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x00010203 &&\n q.lo1 = 0x04050607 &&\n q.hi2 = 0x08090A0B &&\n q.hi3 = 0x0C0D0E0F\n\nlet is_high_dup_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet is_lower_upper_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x04050607 &&\n q.lo1 = 0x00010203 &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet eval_Pshufb_raw (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if is_full_byte_reversal_mask src2 then\n Some (reverse_bytes_quad32 src1)\n else if is_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3))\n else if is_high_dup_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else if is_lower_upper_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else None\n\nlet eval_Pshufb (src1 src2:quad32) : option quad32 = check_ssse3 (eval_Pshufb_raw src1 src2)\nval ins_Pshufb : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pshufb\n\nlet eval_VPshufb (src1 src2:quad32) : option quad32 = check_avx (eval_Pshufb_raw src1 src2)\nval ins_VPshufb : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPshufb\n\nlet eval_Pshufd (permutation:nat8) (src:quad32) : option quad32 = check_sse2 (\n let bits:bits_of_byte = byte_to_twobits permutation in\n Some (Mkfour\n (select_word src bits.lo0)\n (select_word src bits.lo1)\n (select_word src bits.hi2)\n (select_word src bits.hi3)))\nval ins_Pshufd (permutation:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_Pshufd permutation)\n\nlet eval_Pcmpeqd (dst src:quad32) : option quad32 = check_sse2 (\n let eq_result (b:bool):nat32 = if b then 0xFFFFFFFF else 0 in\n Some (Mkfour\n (eq_result (src.lo0 = dst.lo0))\n (eq_result (src.lo1 = dst.lo1))\n (eq_result (src.hi2 = dst.hi2))\n (eq_result (src.hi3 = dst.hi3))))\nval ins_Pcmpeqd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pcmpeqd\n\nlet eval_Pextrq (index:nat8) (src:quad32) : option nat64 = check_sse4_1 (\n let src_two = four_to_two_two src in\n Some (two_to_nat 32 (two_select src_two (index % 2))))\nval ins_Pextrq (index:nat8) : instr_dep [out op64] [opXmm] PreserveFlags (eval_Pextrq index)\n\nlet eval_Pinsrd (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat32 dst (src % pow2_32) (index % 4)))\nval ins_Pinsrd (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrd index)\n\nlet eval_Pinsrq (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat64_def dst src (index % 2)))\nval ins_Pinsrq (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrq index)\n\nlet eval_Pslldq_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the two very special cases we need\n if count = 4 then Some (Mkfour 0 src.lo0 src.lo1 src.hi2)\n else if count = 8 then Some (Mkfour 0 0 src.lo0 src.lo1)\n else None\nlet eval_VPslldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Pslldq_raw count src)\nval ins_VPslldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPslldq count)\n\nlet eval_Psrldq_8_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the one very special case we need\n if count = 8 then Some (Mkfour src.hi2 src.hi3 0 0)\n else None\nlet eval_VPsrldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Psrldq_8_raw count src)\nval ins_VPsrldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPsrldq count)\n\nlet eval_Pclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n let Mkfour a0 a1 a2 a3 = src1 in\n let Mkfour b0 b1 b2 b3 = src2 in\n let f x0 x1 y0 y1 =\n let x = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo x0 x1) in\n let y = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo y0 y1) in\n Vale.Math.Poly2.Bits_s.to_quad32 (Vale.Math.Poly2_s.mul x y)\n in\n if pclmulqdq_enabled then\n match imm with\n | 0 -> Some (f a0 a1 b0 b1)\n | 1 -> Some (f a2 a3 b0 b1)\n | 16 -> Some (f a0 a1 b2 b3)\n | 17 -> Some (f a2 a3 b2 b3)\n | _ -> None\n else None\nval ins_Pclmulqdq (imm:int) : instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Pclmulqdq imm)\n\nlet eval_VPclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Pclmulqdq imm src1 src2)\nval ins_VPclmulqdq (imm:int) : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPclmulqdq imm)\n\nlet eval_AESNI_enc (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.mix_columns_LE (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1))) src2)\n else None\nval ins_AESNI_enc : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc\n\nlet eval_VAESNI_enc (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc src1 src2)\nval ins_VAESNI_enc : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc\n\nlet eval_AESNI_enc_last (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1)) src2)\n else None\nval ins_AESNI_enc_last : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc_last\n\nlet eval_VAESNI_enc_last (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc_last src1 src2)\nval ins_VAESNI_enc_last : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc_last\n\nlet eval_AESNI_dec (dst src:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.inv_mix_columns_LE (Vale.AES.AES_s.inv_sub_bytes (Vale.AES.AES_s.inv_shift_rows_LE dst))) src)\n else None\nval ins_AESNI_dec : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_dec\n\nlet eval_AESNI_dec_last (dst src:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.inv_sub_bytes (Vale.AES.AES_s.inv_shift_rows_LE dst)) src)\n else None\nval ins_AESNI_dec_last : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_dec_last\n\nlet eval_AESNI_imc (src:quad32) : option quad32 =\n if aesni_enabled then Some (Vale.AES.AES_s.inv_mix_columns_LE src) else None\nval ins_AESNI_imc : instr_dep [out opXmm] [opXmm] PreserveFlags eval_AESNI_imc\n\nlet eval_AESNI_keygen_assist (imm:nat8) (src:quad32) : option quad32 =\n if aesni_enabled then\n Some (Mkfour\n (Vale.AES.AES_s.sub_word src.lo1)\n (ixor (Vale.AES.AES_s.rot_word_LE (Vale.AES.AES_s.sub_word src.lo1)) imm)\n (Vale.AES.AES_s.sub_word src.hi3)\n (ixor (Vale.AES.AES_s.rot_word_LE (Vale.AES.AES_s.sub_word src.hi3)) imm))\n else None\nval ins_AESNI_keygen_assist (imm:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_AESNI_keygen_assist imm)\n\nlet eval_SHA256_rnds2 (src1 src2 wk:quad32) : option quad32 =\n if sha_enabled then Some (sha256_rnds2_spec src1 src2 wk) else None\nval ins_SHA256_rnds2 :\n instr_dep [inOut opXmm] [opXmm; oneXmm (OReg 0)] PreserveFlags eval_SHA256_rnds2\n\nlet eval_SHA256_msg1 (src1 src2:quad32) : option quad32 =\n if sha_enabled then Some (sha256_msg1_spec src1 src2) else None\nval ins_SHA256_msg1 : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_SHA256_msg1\n\nlet eval_SHA256_msg2 (src1 src2:quad32) : option quad32 =\n if sha_enabled then Some (sha256_msg2_spec src1 src2) else None\nval ins_SHA256_msg2 : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_SHA256_msg2\n\nPartial implementation:\nlet eval_Ghost:option unit =\n \n"
},
{
"role": "assistant",
"content": "Some ()"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: src1: Vale.X64.Machine_s.quad32 -> src2: Vale.X64.Machine_s.quad32\n -> FStar.Pervasives.Native.option Vale.X64.Machine_s.quad32\nEffect: Prims.Tot\n\nContext:\nmodule Vale.X64.Instructions_s\nopen FStar.Mul\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Two_s\nopen Vale.Def.Words.Four_s\nopen Vale.Def.Types_s\nopen Vale.X64.Machine_s\nopen Vale.X64.Instruction_s\nopen Vale.X64.CPU_Features_s\nopen Vale.X64.CryptoInstructions_s\nopen FStar.Seq.Base\n\nlet eval_Mov64 (src:nat64) : option nat64 = Some src\nval ins_Mov64 : instr_dep [out op64] [op64] PreserveFlags eval_Mov64\n\nlet eval_MovBe64 (src:nat64) : option nat64 = if movbe_enabled then Some (reverse_bytes_nat64 src) else None\nval ins_MovBe64 : instr_dep [out op64] [op64] PreserveFlags eval_MovBe64\n\nlet eval_Bswap64 (dst:nat64) : option nat64 = Some (reverse_bytes_nat64 dst)\nval ins_Bswap64 : instr_dep [inOut op64] [] PreserveFlags eval_Bswap64\n\nlet eval_Cmovc64 (dst src:nat64) (carry:bool) : option nat64 = Some (if carry then src else dst)\nval ins_Cmovc64 : instr_dep [inOut op64] [op64; opFlagsCf] PreserveFlags eval_Cmovc64\n\nlet eval_Add64 (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src in Some (sum >= pow2_64, sum % pow2_64)\nval ins_Add64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Add64\n\nlet eval_AddLea64 (src1 src2:nat64) : option nat64 = Some ((src1 + src2) % pow2_64)\nval ins_AddLea64 :instr_dep [out op64] [op64; op64] PreserveFlags eval_AddLea64\n\nlet eval_AddCarry64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_carry then 1 else 0) in\n Some (sum >= pow2_64, sum % pow2_64)\nval ins_AddCarry64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_AddCarry64\n\nlet eval_Adcx64_Adox64 (old_flag:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_flag then 1 else 0) in\n if adx_enabled then Some (sum >= pow2_64, sum % pow2_64) else None\nval ins_Adcx64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\nval ins_Adox64 : instr_dep [inOut opFlagsOf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\n\nlet eval_Sub64 (dst src:nat64) : option (bool & nat64) =\n let diff = dst - src in Some (diff < 0, diff % pow2_64)\nval ins_Sub64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Sub64\n\nlet eval_Sbb64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let diff = dst - (src + (if old_carry then 1 else 0)) in\n Some (diff < 0, diff % pow2_64) // We specify cf, but underspecify everything else (via HavocFlags)\nval ins_Sbb64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_Sbb64\n\nlet eval_Mul64 (rax src:nat64) : option (nat64 & nat64) =\n Some (FStar.UInt.mul_div #64 rax src, FStar.UInt.mul_mod #64 rax src)\nval ins_Mul64 : instr_dep [out (one64Reg rRdx); inOut (one64Reg rRax)] [op64] HavocFlags eval_Mul64\n\nlet eval_Mulx64 (rdx src:nat64) : option (nat64 & nat64) =\n let hi = FStar.UInt.mul_div #64 rdx src in\n let lo = FStar.UInt.mul_mod #64 rdx src in\n if bmi2_enabled then Some (hi, lo) else None\nval ins_Mulx64 : instr_dep [out op64; out op64] [one64Reg rRdx; op64] PreserveFlags eval_Mulx64\n\nlet eval_IMul64 (dst src:nat64) : option nat64 =\n Some (FStar.UInt.mul_mod #64 dst src)\nval ins_IMul64 : instr_dep [inOut op64] [op64] HavocFlags eval_IMul64\n\nlet eval_And64 (dst src:nat64) : option nat64 = Some (iand dst src)\nval ins_And64 : instr_dep [inOut op64] [op64] HavocFlags eval_And64\n\nlet eval_Xor64 (dst src:nat64) : option (nat64 & (bool & bool)) =\n Some (Vale.Def.Types_s.ixor dst src, (false, false))\nval ins_Xor64 : instr_dep [inOut op64; out opFlagsCf; out opFlagsOf] [op64] HavocFlags eval_Xor64\n\nlet eval_Shr64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishr dst amt) else None\nval ins_Shr64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shr64\n\nlet eval_Shl64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishl dst amt) else None\nval ins_Shl64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shl64\n\nlet eval_Cpuid (rax rcx:nat64) : option (nat64 & (nat64 & (nat64 & nat64))) =\n Some (cpuid rRax rax rcx, (cpuid rRbx rax rcx, (cpuid rRcx rax rcx, cpuid rRdx rax rcx)))\nval ins_Cpuid :\n instr_dep\n [inOut (one64Reg rRax); out (one64Reg rRbx); inOut (one64Reg rRcx); out (one64Reg rRdx)]\n [] PreserveFlags eval_Cpuid\n\n// The XGETBV instruction requires that OSXSAVE (in CPUID) is enabled.\n// We underspecify XGETBV here to only support fetching XCR0, which\n// is supported on any processor supporting the XGETBV instruction\nlet eval_Xgetbv (rcx:nat64) : option (nat64 & nat64) =\n if osxsave_enabled && rcx = 0 then Some (xgetbv rRax rcx, xgetbv rRdx rcx) else None\nval ins_Xgetbv :\n instr_dep\n [out (one64Reg rRax); out (one64Reg rRdx)]\n [one64Reg rRcx] PreserveFlags eval_Xgetbv\n\nlet check_avx (#a:Type0) (x:option a) : option a =\n if avx_enabled then x else None\n\nlet check_sse2 (#a:Type0) (x:option a) : option a =\n if sse2_enabled then x else None\n\nlet check_ssse3 (#a:Type0) (x:option a) : option a =\n if ssse3_enabled then x else None\n\nlet check_sse4_1 (#a:Type0) (x:option a) : option a =\n if sse4_1_enabled then x else None\n\nlet eval_Movdqu (src:quad32) : option quad32 = check_sse2 (Some src)\nval ins_Movdqu : instr_dep [out opXmm] [opXmm] PreserveFlags eval_Movdqu\n\nlet eval_Pxor (dst src:quad32) : option quad32 = check_sse2 (Some (quad32_xor dst src))\nval ins_Pxor : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pxor\n\nlet eval_VPxor (src1 src2:quad32) : option quad32 = check_avx (Some (quad32_xor src1 src2))\nval ins_VPxor : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPxor\n\nlet eval_Pand (dst src:quad32) : option quad32 = check_sse2 (Some (four_map2 (fun di si -> iand di si) dst src))\nval ins_Pand : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pand\n\nlet eval_Paddd_raw (src1 src2:quad32) : option quad32 =\n Some (Mkfour\n ((src1.lo0 + src2.lo0) % pow2_32)\n ((src1.lo1 + src2.lo1) % pow2_32)\n ((src1.hi2 + src2.hi2) % pow2_32)\n ((src1.hi3 + src2.hi3) % pow2_32))\n\nlet eval_Paddd (src1 src2:quad32) : option quad32 = check_sse2 (eval_Paddd_raw src1 src2)\nval ins_Paddd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Paddd\n\nlet eval_VPaddd (src1 src2:quad32) : option quad32 = check_avx (eval_Paddd_raw src1 src2)\nval ins_VPaddd : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPaddd\n\nlet eval_Pslld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishl i amt) dst) else None)\nval ins_Pslld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Pslld amt)\n\nlet eval_Psrld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishr i amt) dst) else None)\nval ins_Psrld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrld amt)\n\nlet eval_Psrldq (amt:int) (dst:quad32) : option quad32 = check_sse2 (\n if 0 <= amt && amt < 16 then\n let src_bytes = le_quad32_to_bytes dst in\n let zero_pad = Seq.create amt 0 in\n let remaining_bytes = slice src_bytes amt (length src_bytes) in\n Some (le_bytes_to_quad32 (append zero_pad remaining_bytes))\n else None)\nval ins_Psrldq (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrldq amt)\n\nlet eval_Palignr_raw (amount:nat8) (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if amount = 4 then\n Some (Mkfour src2.lo1 src2.hi2 src2.hi3 src1.lo0)\n else if amount = 8 then\n Some (Mkfour src2.hi2 src2.hi3 src1.lo0 src1.lo1)\n else None\n\nlet eval_Palignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_ssse3 (eval_Palignr_raw amount src1 src2)\nval ins_Palignr (amount:nat8) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Palignr amount)\n\nlet eval_VPalignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Palignr_raw amount src1 src2)\nval ins_VPalignr (amount:nat8) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPalignr amount)\n\nlet eval_Shufpd_raw (permutation:int) (src1 src2:quad32) : option quad32 =\n if 0 <= permutation && permutation < 4 then\n Some (Mkfour\n (if permutation % 2 = 0 then src1.lo0 else src1.hi2)\n (if permutation % 2 = 0 then src1.lo1 else src1.hi3)\n (if (permutation / 2) % 2 = 0 then src2.lo0 else src2.hi2)\n (if (permutation / 2) % 2 = 0 then src2.lo1 else src2.hi3))\n else None\n\nlet eval_Shufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_sse2 (eval_Shufpd_raw permutation src1 src2)\nval ins_Shufpd (permutation:int) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Shufpd permutation)\n\nlet eval_VShufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Shufpd_raw permutation src1 src2)\nval ins_VShufpd (permutation:int) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VShufpd permutation)\n\nlet is_full_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x04050607 &&\n q.hi3 = 0x00010203\n\nlet is_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x00010203 &&\n q.lo1 = 0x04050607 &&\n q.hi2 = 0x08090A0B &&\n q.hi3 = 0x0C0D0E0F\n\nlet is_high_dup_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet is_lower_upper_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x04050607 &&\n q.lo1 = 0x00010203 &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet eval_Pshufb_raw (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if is_full_byte_reversal_mask src2 then\n Some (reverse_bytes_quad32 src1)\n else if is_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3))\n else if is_high_dup_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else if is_lower_upper_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else None\n\nlet eval_Pshufb (src1 src2:quad32) : option quad32 = check_ssse3 (eval_Pshufb_raw src1 src2)\nval ins_Pshufb : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pshufb\n\nlet eval_VPshufb (src1 src2:quad32) : option quad32 = check_avx (eval_Pshufb_raw src1 src2)\nval ins_VPshufb : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPshufb\n\nlet eval_Pshufd (permutation:nat8) (src:quad32) : option quad32 = check_sse2 (\n let bits:bits_of_byte = byte_to_twobits permutation in\n Some (Mkfour\n (select_word src bits.lo0)\n (select_word src bits.lo1)\n (select_word src bits.hi2)\n (select_word src bits.hi3)))\nval ins_Pshufd (permutation:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_Pshufd permutation)\n\nlet eval_Pcmpeqd (dst src:quad32) : option quad32 = check_sse2 (\n let eq_result (b:bool):nat32 = if b then 0xFFFFFFFF else 0 in\n Some (Mkfour\n (eq_result (src.lo0 = dst.lo0))\n (eq_result (src.lo1 = dst.lo1))\n (eq_result (src.hi2 = dst.hi2))\n (eq_result (src.hi3 = dst.hi3))))\nval ins_Pcmpeqd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pcmpeqd\n\nlet eval_Pextrq (index:nat8) (src:quad32) : option nat64 = check_sse4_1 (\n let src_two = four_to_two_two src in\n Some (two_to_nat 32 (two_select src_two (index % 2))))\nval ins_Pextrq (index:nat8) : instr_dep [out op64] [opXmm] PreserveFlags (eval_Pextrq index)\n\nlet eval_Pinsrd (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat32 dst (src % pow2_32) (index % 4)))\nval ins_Pinsrd (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrd index)\n\nlet eval_Pinsrq (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat64_def dst src (index % 2)))\nval ins_Pinsrq (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrq index)\n\nlet eval_Pslldq_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the two very special cases we need\n if count = 4 then Some (Mkfour 0 src.lo0 src.lo1 src.hi2)\n else if count = 8 then Some (Mkfour 0 0 src.lo0 src.lo1)\n else None\nlet eval_VPslldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Pslldq_raw count src)\nval ins_VPslldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPslldq count)\n\nlet eval_Psrldq_8_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the one very special case we need\n if count = 8 then Some (Mkfour src.hi2 src.hi3 0 0)\n else None\nlet eval_VPsrldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Psrldq_8_raw count src)\nval ins_VPsrldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPsrldq count)\n\nlet eval_Pclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n let Mkfour a0 a1 a2 a3 = src1 in\n let Mkfour b0 b1 b2 b3 = src2 in\n let f x0 x1 y0 y1 =\n let x = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo x0 x1) in\n let y = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo y0 y1) in\n Vale.Math.Poly2.Bits_s.to_quad32 (Vale.Math.Poly2_s.mul x y)\n in\n if pclmulqdq_enabled then\n match imm with\n | 0 -> Some (f a0 a1 b0 b1)\n | 1 -> Some (f a2 a3 b0 b1)\n | 16 -> Some (f a0 a1 b2 b3)\n | 17 -> Some (f a2 a3 b2 b3)\n | _ -> None\n else None\nval ins_Pclmulqdq (imm:int) : instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Pclmulqdq imm)\n\nlet eval_VPclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Pclmulqdq imm src1 src2)\nval ins_VPclmulqdq (imm:int) : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPclmulqdq imm)\n\nlet eval_AESNI_enc (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.mix_columns_LE (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1))) src2)\n else None\nval ins_AESNI_enc : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc\n\nlet eval_VAESNI_enc (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc src1 src2)\nval ins_VAESNI_enc : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc\n\nlet eval_AESNI_enc_last (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1)) src2)\n else None\nval ins_AESNI_enc_last : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc_last\n\nPartial implementation:\nlet eval_VAESNI_enc_last (src1 src2: quad32) : option quad32 =\n \n"
},
{
"role": "assistant",
"content": "check_avx (eval_AESNI_enc_last src1 src2)"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: count: Vale.Def.Types_s.nat8 -> src: Vale.X64.Machine_s.quad32\n -> FStar.Pervasives.Native.option Vale.X64.Machine_s.quad32\nEffect: Prims.Tot\n\nContext:\nmodule Vale.X64.Instructions_s\nopen FStar.Mul\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Two_s\nopen Vale.Def.Words.Four_s\nopen Vale.Def.Types_s\nopen Vale.X64.Machine_s\nopen Vale.X64.Instruction_s\nopen Vale.X64.CPU_Features_s\nopen Vale.X64.CryptoInstructions_s\nopen FStar.Seq.Base\n\nlet eval_Mov64 (src:nat64) : option nat64 = Some src\nval ins_Mov64 : instr_dep [out op64] [op64] PreserveFlags eval_Mov64\n\nlet eval_MovBe64 (src:nat64) : option nat64 = if movbe_enabled then Some (reverse_bytes_nat64 src) else None\nval ins_MovBe64 : instr_dep [out op64] [op64] PreserveFlags eval_MovBe64\n\nlet eval_Bswap64 (dst:nat64) : option nat64 = Some (reverse_bytes_nat64 dst)\nval ins_Bswap64 : instr_dep [inOut op64] [] PreserveFlags eval_Bswap64\n\nlet eval_Cmovc64 (dst src:nat64) (carry:bool) : option nat64 = Some (if carry then src else dst)\nval ins_Cmovc64 : instr_dep [inOut op64] [op64; opFlagsCf] PreserveFlags eval_Cmovc64\n\nlet eval_Add64 (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src in Some (sum >= pow2_64, sum % pow2_64)\nval ins_Add64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Add64\n\nlet eval_AddLea64 (src1 src2:nat64) : option nat64 = Some ((src1 + src2) % pow2_64)\nval ins_AddLea64 :instr_dep [out op64] [op64; op64] PreserveFlags eval_AddLea64\n\nlet eval_AddCarry64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_carry then 1 else 0) in\n Some (sum >= pow2_64, sum % pow2_64)\nval ins_AddCarry64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_AddCarry64\n\nlet eval_Adcx64_Adox64 (old_flag:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_flag then 1 else 0) in\n if adx_enabled then Some (sum >= pow2_64, sum % pow2_64) else None\nval ins_Adcx64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\nval ins_Adox64 : instr_dep [inOut opFlagsOf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\n\nlet eval_Sub64 (dst src:nat64) : option (bool & nat64) =\n let diff = dst - src in Some (diff < 0, diff % pow2_64)\nval ins_Sub64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Sub64\n\nlet eval_Sbb64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let diff = dst - (src + (if old_carry then 1 else 0)) in\n Some (diff < 0, diff % pow2_64) // We specify cf, but underspecify everything else (via HavocFlags)\nval ins_Sbb64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_Sbb64\n\nlet eval_Mul64 (rax src:nat64) : option (nat64 & nat64) =\n Some (FStar.UInt.mul_div #64 rax src, FStar.UInt.mul_mod #64 rax src)\nval ins_Mul64 : instr_dep [out (one64Reg rRdx); inOut (one64Reg rRax)] [op64] HavocFlags eval_Mul64\n\nlet eval_Mulx64 (rdx src:nat64) : option (nat64 & nat64) =\n let hi = FStar.UInt.mul_div #64 rdx src in\n let lo = FStar.UInt.mul_mod #64 rdx src in\n if bmi2_enabled then Some (hi, lo) else None\nval ins_Mulx64 : instr_dep [out op64; out op64] [one64Reg rRdx; op64] PreserveFlags eval_Mulx64\n\nlet eval_IMul64 (dst src:nat64) : option nat64 =\n Some (FStar.UInt.mul_mod #64 dst src)\nval ins_IMul64 : instr_dep [inOut op64] [op64] HavocFlags eval_IMul64\n\nlet eval_And64 (dst src:nat64) : option nat64 = Some (iand dst src)\nval ins_And64 : instr_dep [inOut op64] [op64] HavocFlags eval_And64\n\nlet eval_Xor64 (dst src:nat64) : option (nat64 & (bool & bool)) =\n Some (Vale.Def.Types_s.ixor dst src, (false, false))\nval ins_Xor64 : instr_dep [inOut op64; out opFlagsCf; out opFlagsOf] [op64] HavocFlags eval_Xor64\n\nlet eval_Shr64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishr dst amt) else None\nval ins_Shr64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shr64\n\nlet eval_Shl64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishl dst amt) else None\nval ins_Shl64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shl64\n\nlet eval_Cpuid (rax rcx:nat64) : option (nat64 & (nat64 & (nat64 & nat64))) =\n Some (cpuid rRax rax rcx, (cpuid rRbx rax rcx, (cpuid rRcx rax rcx, cpuid rRdx rax rcx)))\nval ins_Cpuid :\n instr_dep\n [inOut (one64Reg rRax); out (one64Reg rRbx); inOut (one64Reg rRcx); out (one64Reg rRdx)]\n [] PreserveFlags eval_Cpuid\n\n// The XGETBV instruction requires that OSXSAVE (in CPUID) is enabled.\n// We underspecify XGETBV here to only support fetching XCR0, which\n// is supported on any processor supporting the XGETBV instruction\nlet eval_Xgetbv (rcx:nat64) : option (nat64 & nat64) =\n if osxsave_enabled && rcx = 0 then Some (xgetbv rRax rcx, xgetbv rRdx rcx) else None\nval ins_Xgetbv :\n instr_dep\n [out (one64Reg rRax); out (one64Reg rRdx)]\n [one64Reg rRcx] PreserveFlags eval_Xgetbv\n\nlet check_avx (#a:Type0) (x:option a) : option a =\n if avx_enabled then x else None\n\nlet check_sse2 (#a:Type0) (x:option a) : option a =\n if sse2_enabled then x else None\n\nlet check_ssse3 (#a:Type0) (x:option a) : option a =\n if ssse3_enabled then x else None\n\nlet check_sse4_1 (#a:Type0) (x:option a) : option a =\n if sse4_1_enabled then x else None\n\nlet eval_Movdqu (src:quad32) : option quad32 = check_sse2 (Some src)\nval ins_Movdqu : instr_dep [out opXmm] [opXmm] PreserveFlags eval_Movdqu\n\nlet eval_Pxor (dst src:quad32) : option quad32 = check_sse2 (Some (quad32_xor dst src))\nval ins_Pxor : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pxor\n\nlet eval_VPxor (src1 src2:quad32) : option quad32 = check_avx (Some (quad32_xor src1 src2))\nval ins_VPxor : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPxor\n\nlet eval_Pand (dst src:quad32) : option quad32 = check_sse2 (Some (four_map2 (fun di si -> iand di si) dst src))\nval ins_Pand : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pand\n\nlet eval_Paddd_raw (src1 src2:quad32) : option quad32 =\n Some (Mkfour\n ((src1.lo0 + src2.lo0) % pow2_32)\n ((src1.lo1 + src2.lo1) % pow2_32)\n ((src1.hi2 + src2.hi2) % pow2_32)\n ((src1.hi3 + src2.hi3) % pow2_32))\n\nlet eval_Paddd (src1 src2:quad32) : option quad32 = check_sse2 (eval_Paddd_raw src1 src2)\nval ins_Paddd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Paddd\n\nlet eval_VPaddd (src1 src2:quad32) : option quad32 = check_avx (eval_Paddd_raw src1 src2)\nval ins_VPaddd : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPaddd\n\nlet eval_Pslld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishl i amt) dst) else None)\nval ins_Pslld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Pslld amt)\n\nlet eval_Psrld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishr i amt) dst) else None)\nval ins_Psrld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrld amt)\n\nlet eval_Psrldq (amt:int) (dst:quad32) : option quad32 = check_sse2 (\n if 0 <= amt && amt < 16 then\n let src_bytes = le_quad32_to_bytes dst in\n let zero_pad = Seq.create amt 0 in\n let remaining_bytes = slice src_bytes amt (length src_bytes) in\n Some (le_bytes_to_quad32 (append zero_pad remaining_bytes))\n else None)\nval ins_Psrldq (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrldq amt)\n\nlet eval_Palignr_raw (amount:nat8) (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if amount = 4 then\n Some (Mkfour src2.lo1 src2.hi2 src2.hi3 src1.lo0)\n else if amount = 8 then\n Some (Mkfour src2.hi2 src2.hi3 src1.lo0 src1.lo1)\n else None\n\nlet eval_Palignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_ssse3 (eval_Palignr_raw amount src1 src2)\nval ins_Palignr (amount:nat8) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Palignr amount)\n\nlet eval_VPalignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Palignr_raw amount src1 src2)\nval ins_VPalignr (amount:nat8) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPalignr amount)\n\nlet eval_Shufpd_raw (permutation:int) (src1 src2:quad32) : option quad32 =\n if 0 <= permutation && permutation < 4 then\n Some (Mkfour\n (if permutation % 2 = 0 then src1.lo0 else src1.hi2)\n (if permutation % 2 = 0 then src1.lo1 else src1.hi3)\n (if (permutation / 2) % 2 = 0 then src2.lo0 else src2.hi2)\n (if (permutation / 2) % 2 = 0 then src2.lo1 else src2.hi3))\n else None\n\nlet eval_Shufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_sse2 (eval_Shufpd_raw permutation src1 src2)\nval ins_Shufpd (permutation:int) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Shufpd permutation)\n\nlet eval_VShufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Shufpd_raw permutation src1 src2)\nval ins_VShufpd (permutation:int) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VShufpd permutation)\n\nlet is_full_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x04050607 &&\n q.hi3 = 0x00010203\n\nlet is_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x00010203 &&\n q.lo1 = 0x04050607 &&\n q.hi2 = 0x08090A0B &&\n q.hi3 = 0x0C0D0E0F\n\nlet is_high_dup_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet is_lower_upper_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x04050607 &&\n q.lo1 = 0x00010203 &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet eval_Pshufb_raw (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if is_full_byte_reversal_mask src2 then\n Some (reverse_bytes_quad32 src1)\n else if is_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3))\n else if is_high_dup_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else if is_lower_upper_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else None\n\nlet eval_Pshufb (src1 src2:quad32) : option quad32 = check_ssse3 (eval_Pshufb_raw src1 src2)\nval ins_Pshufb : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pshufb\n\nlet eval_VPshufb (src1 src2:quad32) : option quad32 = check_avx (eval_Pshufb_raw src1 src2)\nval ins_VPshufb : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPshufb\n\nlet eval_Pshufd (permutation:nat8) (src:quad32) : option quad32 = check_sse2 (\n let bits:bits_of_byte = byte_to_twobits permutation in\n Some (Mkfour\n (select_word src bits.lo0)\n (select_word src bits.lo1)\n (select_word src bits.hi2)\n (select_word src bits.hi3)))\nval ins_Pshufd (permutation:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_Pshufd permutation)\n\nlet eval_Pcmpeqd (dst src:quad32) : option quad32 = check_sse2 (\n let eq_result (b:bool):nat32 = if b then 0xFFFFFFFF else 0 in\n Some (Mkfour\n (eq_result (src.lo0 = dst.lo0))\n (eq_result (src.lo1 = dst.lo1))\n (eq_result (src.hi2 = dst.hi2))\n (eq_result (src.hi3 = dst.hi3))))\nval ins_Pcmpeqd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pcmpeqd\n\nlet eval_Pextrq (index:nat8) (src:quad32) : option nat64 = check_sse4_1 (\n let src_two = four_to_two_two src in\n Some (two_to_nat 32 (two_select src_two (index % 2))))\nval ins_Pextrq (index:nat8) : instr_dep [out op64] [opXmm] PreserveFlags (eval_Pextrq index)\n\nlet eval_Pinsrd (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat32 dst (src % pow2_32) (index % 4)))\nval ins_Pinsrd (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrd index)\n\nlet eval_Pinsrq (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat64_def dst src (index % 2)))\nval ins_Pinsrq (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrq index)\n\nlet eval_Pslldq_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the two very special cases we need\n if count = 4 then Some (Mkfour 0 src.lo0 src.lo1 src.hi2)\n else if count = 8 then Some (Mkfour 0 0 src.lo0 src.lo1)\n else None\nlet eval_VPslldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Pslldq_raw count src)\nval ins_VPslldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPslldq count)\n\nlet eval_Psrldq_8_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the one very special case we need\n if count = 8 then Some (Mkfour src.hi2 src.hi3 0 0)\n\nPartial implementation:\nlet eval_VPsrldq (count: nat8) (src: quad32) : option quad32 =\n \n"
},
{
"role": "assistant",
"content": "check_avx (eval_Psrldq_8_raw count src)"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: src1: Vale.X64.Machine_s.quad32 -> src2: Vale.X64.Machine_s.quad32\n -> FStar.Pervasives.Native.option Vale.X64.Machine_s.quad32\nEffect: Prims.Tot\n\nContext:\nmodule Vale.X64.Instructions_s\nopen FStar.Mul\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Two_s\nopen Vale.Def.Words.Four_s\nopen Vale.Def.Types_s\nopen Vale.X64.Machine_s\nopen Vale.X64.Instruction_s\nopen Vale.X64.CPU_Features_s\nopen Vale.X64.CryptoInstructions_s\nopen FStar.Seq.Base\n\nlet eval_Mov64 (src:nat64) : option nat64 = Some src\nval ins_Mov64 : instr_dep [out op64] [op64] PreserveFlags eval_Mov64\n\nlet eval_MovBe64 (src:nat64) : option nat64 = if movbe_enabled then Some (reverse_bytes_nat64 src) else None\nval ins_MovBe64 : instr_dep [out op64] [op64] PreserveFlags eval_MovBe64\n\nlet eval_Bswap64 (dst:nat64) : option nat64 = Some (reverse_bytes_nat64 dst)\nval ins_Bswap64 : instr_dep [inOut op64] [] PreserveFlags eval_Bswap64\n\nlet eval_Cmovc64 (dst src:nat64) (carry:bool) : option nat64 = Some (if carry then src else dst)\nval ins_Cmovc64 : instr_dep [inOut op64] [op64; opFlagsCf] PreserveFlags eval_Cmovc64\n\nlet eval_Add64 (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src in Some (sum >= pow2_64, sum % pow2_64)\nval ins_Add64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Add64\n\nlet eval_AddLea64 (src1 src2:nat64) : option nat64 = Some ((src1 + src2) % pow2_64)\nval ins_AddLea64 :instr_dep [out op64] [op64; op64] PreserveFlags eval_AddLea64\n\nlet eval_AddCarry64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_carry then 1 else 0) in\n Some (sum >= pow2_64, sum % pow2_64)\nval ins_AddCarry64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_AddCarry64\n\nlet eval_Adcx64_Adox64 (old_flag:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_flag then 1 else 0) in\n if adx_enabled then Some (sum >= pow2_64, sum % pow2_64) else None\nval ins_Adcx64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\nval ins_Adox64 : instr_dep [inOut opFlagsOf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\n\nlet eval_Sub64 (dst src:nat64) : option (bool & nat64) =\n let diff = dst - src in Some (diff < 0, diff % pow2_64)\nval ins_Sub64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Sub64\n\nlet eval_Sbb64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let diff = dst - (src + (if old_carry then 1 else 0)) in\n Some (diff < 0, diff % pow2_64) // We specify cf, but underspecify everything else (via HavocFlags)\nval ins_Sbb64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_Sbb64\n\nlet eval_Mul64 (rax src:nat64) : option (nat64 & nat64) =\n Some (FStar.UInt.mul_div #64 rax src, FStar.UInt.mul_mod #64 rax src)\nval ins_Mul64 : instr_dep [out (one64Reg rRdx); inOut (one64Reg rRax)] [op64] HavocFlags eval_Mul64\n\nlet eval_Mulx64 (rdx src:nat64) : option (nat64 & nat64) =\n let hi = FStar.UInt.mul_div #64 rdx src in\n let lo = FStar.UInt.mul_mod #64 rdx src in\n if bmi2_enabled then Some (hi, lo) else None\nval ins_Mulx64 : instr_dep [out op64; out op64] [one64Reg rRdx; op64] PreserveFlags eval_Mulx64\n\nlet eval_IMul64 (dst src:nat64) : option nat64 =\n Some (FStar.UInt.mul_mod #64 dst src)\nval ins_IMul64 : instr_dep [inOut op64] [op64] HavocFlags eval_IMul64\n\nlet eval_And64 (dst src:nat64) : option nat64 = Some (iand dst src)\nval ins_And64 : instr_dep [inOut op64] [op64] HavocFlags eval_And64\n\nlet eval_Xor64 (dst src:nat64) : option (nat64 & (bool & bool)) =\n Some (Vale.Def.Types_s.ixor dst src, (false, false))\nval ins_Xor64 : instr_dep [inOut op64; out opFlagsCf; out opFlagsOf] [op64] HavocFlags eval_Xor64\n\nlet eval_Shr64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishr dst amt) else None\nval ins_Shr64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shr64\n\nlet eval_Shl64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishl dst amt) else None\nval ins_Shl64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shl64\n\nlet eval_Cpuid (rax rcx:nat64) : option (nat64 & (nat64 & (nat64 & nat64))) =\n Some (cpuid rRax rax rcx, (cpuid rRbx rax rcx, (cpuid rRcx rax rcx, cpuid rRdx rax rcx)))\nval ins_Cpuid :\n instr_dep\n [inOut (one64Reg rRax); out (one64Reg rRbx); inOut (one64Reg rRcx); out (one64Reg rRdx)]\n [] PreserveFlags eval_Cpuid\n\n// The XGETBV instruction requires that OSXSAVE (in CPUID) is enabled.\n// We underspecify XGETBV here to only support fetching XCR0, which\n// is supported on any processor supporting the XGETBV instruction\nlet eval_Xgetbv (rcx:nat64) : option (nat64 & nat64) =\n if osxsave_enabled && rcx = 0 then Some (xgetbv rRax rcx, xgetbv rRdx rcx) else None\nval ins_Xgetbv :\n instr_dep\n [out (one64Reg rRax); out (one64Reg rRdx)]\n [one64Reg rRcx] PreserveFlags eval_Xgetbv\n\nlet check_avx (#a:Type0) (x:option a) : option a =\n if avx_enabled then x else None\n\nlet check_sse2 (#a:Type0) (x:option a) : option a =\n if sse2_enabled then x else None\n\nlet check_ssse3 (#a:Type0) (x:option a) : option a =\n if ssse3_enabled then x else None\n\nlet check_sse4_1 (#a:Type0) (x:option a) : option a =\n if sse4_1_enabled then x else None\n\nlet eval_Movdqu (src:quad32) : option quad32 = check_sse2 (Some src)\nval ins_Movdqu : instr_dep [out opXmm] [opXmm] PreserveFlags eval_Movdqu\n\nlet eval_Pxor (dst src:quad32) : option quad32 = check_sse2 (Some (quad32_xor dst src))\nval ins_Pxor : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pxor\n\nlet eval_VPxor (src1 src2:quad32) : option quad32 = check_avx (Some (quad32_xor src1 src2))\nval ins_VPxor : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPxor\n\nlet eval_Pand (dst src:quad32) : option quad32 = check_sse2 (Some (four_map2 (fun di si -> iand di si) dst src))\nval ins_Pand : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pand\n\nlet eval_Paddd_raw (src1 src2:quad32) : option quad32 =\n Some (Mkfour\n ((src1.lo0 + src2.lo0) % pow2_32)\n ((src1.lo1 + src2.lo1) % pow2_32)\n ((src1.hi2 + src2.hi2) % pow2_32)\n ((src1.hi3 + src2.hi3) % pow2_32))\n\nlet eval_Paddd (src1 src2:quad32) : option quad32 = check_sse2 (eval_Paddd_raw src1 src2)\nval ins_Paddd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Paddd\n\nlet eval_VPaddd (src1 src2:quad32) : option quad32 = check_avx (eval_Paddd_raw src1 src2)\nval ins_VPaddd : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPaddd\n\nlet eval_Pslld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishl i amt) dst) else None)\nval ins_Pslld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Pslld amt)\n\nlet eval_Psrld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishr i amt) dst) else None)\nval ins_Psrld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrld amt)\n\nlet eval_Psrldq (amt:int) (dst:quad32) : option quad32 = check_sse2 (\n if 0 <= amt && amt < 16 then\n let src_bytes = le_quad32_to_bytes dst in\n let zero_pad = Seq.create amt 0 in\n let remaining_bytes = slice src_bytes amt (length src_bytes) in\n Some (le_bytes_to_quad32 (append zero_pad remaining_bytes))\n else None)\nval ins_Psrldq (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrldq amt)\n\nlet eval_Palignr_raw (amount:nat8) (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if amount = 4 then\n Some (Mkfour src2.lo1 src2.hi2 src2.hi3 src1.lo0)\n else if amount = 8 then\n Some (Mkfour src2.hi2 src2.hi3 src1.lo0 src1.lo1)\n else None\n\nlet eval_Palignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_ssse3 (eval_Palignr_raw amount src1 src2)\nval ins_Palignr (amount:nat8) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Palignr amount)\n\nlet eval_VPalignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Palignr_raw amount src1 src2)\nval ins_VPalignr (amount:nat8) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPalignr amount)\n\nlet eval_Shufpd_raw (permutation:int) (src1 src2:quad32) : option quad32 =\n if 0 <= permutation && permutation < 4 then\n Some (Mkfour\n (if permutation % 2 = 0 then src1.lo0 else src1.hi2)\n (if permutation % 2 = 0 then src1.lo1 else src1.hi3)\n (if (permutation / 2) % 2 = 0 then src2.lo0 else src2.hi2)\n (if (permutation / 2) % 2 = 0 then src2.lo1 else src2.hi3))\n else None\n\nlet eval_Shufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_sse2 (eval_Shufpd_raw permutation src1 src2)\nval ins_Shufpd (permutation:int) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Shufpd permutation)\n\nlet eval_VShufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Shufpd_raw permutation src1 src2)\nval ins_VShufpd (permutation:int) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VShufpd permutation)\n\nlet is_full_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x04050607 &&\n q.hi3 = 0x00010203\n\nlet is_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x00010203 &&\n q.lo1 = 0x04050607 &&\n q.hi2 = 0x08090A0B &&\n q.hi3 = 0x0C0D0E0F\n\nlet is_high_dup_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet is_lower_upper_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x04050607 &&\n q.lo1 = 0x00010203 &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet eval_Pshufb_raw (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if is_full_byte_reversal_mask src2 then\n Some (reverse_bytes_quad32 src1)\n else if is_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3))\n else if is_high_dup_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else if is_lower_upper_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else None\n\nlet eval_Pshufb (src1 src2:quad32) : option quad32 = check_ssse3 (eval_Pshufb_raw src1 src2)\nval ins_Pshufb : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pshufb\n\nlet eval_VPshufb (src1 src2:quad32) : option quad32 = check_avx (eval_Pshufb_raw src1 src2)\nval ins_VPshufb : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPshufb\n\nlet eval_Pshufd (permutation:nat8) (src:quad32) : option quad32 = check_sse2 (\n let bits:bits_of_byte = byte_to_twobits permutation in\n Some (Mkfour\n (select_word src bits.lo0)\n (select_word src bits.lo1)\n (select_word src bits.hi2)\n (select_word src bits.hi3)))\nval ins_Pshufd (permutation:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_Pshufd permutation)\n\nlet eval_Pcmpeqd (dst src:quad32) : option quad32 = check_sse2 (\n let eq_result (b:bool):nat32 = if b then 0xFFFFFFFF else 0 in\n Some (Mkfour\n (eq_result (src.lo0 = dst.lo0))\n (eq_result (src.lo1 = dst.lo1))\n (eq_result (src.hi2 = dst.hi2))\n (eq_result (src.hi3 = dst.hi3))))\nval ins_Pcmpeqd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pcmpeqd\n\nlet eval_Pextrq (index:nat8) (src:quad32) : option nat64 = check_sse4_1 (\n let src_two = four_to_two_two src in\n Some (two_to_nat 32 (two_select src_two (index % 2))))\nval ins_Pextrq (index:nat8) : instr_dep [out op64] [opXmm] PreserveFlags (eval_Pextrq index)\n\nlet eval_Pinsrd (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat32 dst (src % pow2_32) (index % 4)))\nval ins_Pinsrd (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrd index)\n\nlet eval_Pinsrq (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat64_def dst src (index % 2)))\nval ins_Pinsrq (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrq index)\n\nlet eval_Pslldq_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the two very special cases we need\n if count = 4 then Some (Mkfour 0 src.lo0 src.lo1 src.hi2)\n else if count = 8 then Some (Mkfour 0 0 src.lo0 src.lo1)\n else None\nlet eval_VPslldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Pslldq_raw count src)\nval ins_VPslldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPslldq count)\n\nlet eval_Psrldq_8_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the one very special case we need\n if count = 8 then Some (Mkfour src.hi2 src.hi3 0 0)\n else None\nlet eval_VPsrldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Psrldq_8_raw count src)\nval ins_VPsrldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPsrldq count)\n\nlet eval_Pclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n let Mkfour a0 a1 a2 a3 = src1 in\n let Mkfour b0 b1 b2 b3 = src2 in\n let f x0 x1 y0 y1 =\n let x = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo x0 x1) in\n let y = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo y0 y1) in\n Vale.Math.Poly2.Bits_s.to_quad32 (Vale.Math.Poly2_s.mul x y)\n in\n if pclmulqdq_enabled then\n match imm with\n | 0 -> Some (f a0 a1 b0 b1)\n | 1 -> Some (f a2 a3 b0 b1)\n | 16 -> Some (f a0 a1 b2 b3)\n | 17 -> Some (f a2 a3 b2 b3)\n | _ -> None\n else None\nval ins_Pclmulqdq (imm:int) : instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Pclmulqdq imm)\n\nlet eval_VPclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Pclmulqdq imm src1 src2)\nval ins_VPclmulqdq (imm:int) : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPclmulqdq imm)\n\nlet eval_AESNI_enc (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.mix_columns_LE (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1))) src2)\n else None\nval ins_AESNI_enc : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc\n\nlet eval_VAESNI_enc (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc src1 src2)\nval ins_VAESNI_enc : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc\n\nlet eval_AESNI_enc_last (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1)) src2)\n else None\nval ins_AESNI_enc_last : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc_last\n\nlet eval_VAESNI_enc_last (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc_last src1 src2)\nval ins_VAESNI_enc_last : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc_last\n\nlet eval_AESNI_dec (dst src:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.inv_mix_columns_LE (Vale.AES.AES_s.inv_sub_bytes (Vale.AES.AES_s.inv_shift_rows_LE dst))) src)\n else None\nval ins_AESNI_dec : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_dec\n\nlet eval_AESNI_dec_last (dst src:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.inv_sub_bytes (Vale.AES.AES_s.inv_shift_rows_LE dst)) src)\n else None\nval ins_AESNI_dec_last : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_dec_last\n\nlet eval_AESNI_imc (src:quad32) : option quad32 =\n if aesni_enabled then Some (Vale.AES.AES_s.inv_mix_columns_LE src) else None\nval ins_AESNI_imc : instr_dep [out opXmm] [opXmm] PreserveFlags eval_AESNI_imc\n\nlet eval_AESNI_keygen_assist (imm:nat8) (src:quad32) : option quad32 =\n if aesni_enabled then\n Some (Mkfour\n (Vale.AES.AES_s.sub_word src.lo1)\n (ixor (Vale.AES.AES_s.rot_word_LE (Vale.AES.AES_s.sub_word src.lo1)) imm)\n (Vale.AES.AES_s.sub_word src.hi3)\n (ixor (Vale.AES.AES_s.rot_word_LE (Vale.AES.AES_s.sub_word src.hi3)) imm))\n else None\nval ins_AESNI_keygen_assist (imm:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_AESNI_keygen_assist imm)\n\nlet eval_SHA256_rnds2 (src1 src2 wk:quad32) : option quad32 =\n if sha_enabled then Some (sha256_rnds2_spec src1 src2 wk) else None\nval ins_SHA256_rnds2 :\n instr_dep [inOut opXmm] [opXmm; oneXmm (OReg 0)] PreserveFlags eval_SHA256_rnds2\n\nlet eval_SHA256_msg1 (src1 src2:quad32) : option quad32 =\n if sha_enabled then Some (sha256_msg1_spec src1 src2) else None\nval ins_SHA256_msg1 : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_SHA256_msg1\n\nPartial implementation:\nlet eval_SHA256_msg2 (src1 src2: quad32) : option quad32 =\n \n"
},
{
"role": "assistant",
"content": "if sha_enabled then Some (sha256_msg2_spec src1 src2) else None"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: FStar.Pervasives.Native.option Prims.unit\nEffect: Prims.Tot\n\nContext:\nmodule Vale.X64.Instructions_s\nopen FStar.Mul\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Two_s\nopen Vale.Def.Words.Four_s\nopen Vale.Def.Types_s\nopen Vale.X64.Machine_s\nopen Vale.X64.Instruction_s\nopen Vale.X64.CPU_Features_s\nopen Vale.X64.CryptoInstructions_s\nopen FStar.Seq.Base\n\nlet eval_Mov64 (src:nat64) : option nat64 = Some src\nval ins_Mov64 : instr_dep [out op64] [op64] PreserveFlags eval_Mov64\n\nlet eval_MovBe64 (src:nat64) : option nat64 = if movbe_enabled then Some (reverse_bytes_nat64 src) else None\nval ins_MovBe64 : instr_dep [out op64] [op64] PreserveFlags eval_MovBe64\n\nlet eval_Bswap64 (dst:nat64) : option nat64 = Some (reverse_bytes_nat64 dst)\nval ins_Bswap64 : instr_dep [inOut op64] [] PreserveFlags eval_Bswap64\n\nlet eval_Cmovc64 (dst src:nat64) (carry:bool) : option nat64 = Some (if carry then src else dst)\nval ins_Cmovc64 : instr_dep [inOut op64] [op64; opFlagsCf] PreserveFlags eval_Cmovc64\n\nlet eval_Add64 (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src in Some (sum >= pow2_64, sum % pow2_64)\nval ins_Add64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Add64\n\nlet eval_AddLea64 (src1 src2:nat64) : option nat64 = Some ((src1 + src2) % pow2_64)\nval ins_AddLea64 :instr_dep [out op64] [op64; op64] PreserveFlags eval_AddLea64\n\nlet eval_AddCarry64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_carry then 1 else 0) in\n Some (sum >= pow2_64, sum % pow2_64)\nval ins_AddCarry64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_AddCarry64\n\nlet eval_Adcx64_Adox64 (old_flag:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_flag then 1 else 0) in\n if adx_enabled then Some (sum >= pow2_64, sum % pow2_64) else None\nval ins_Adcx64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\nval ins_Adox64 : instr_dep [inOut opFlagsOf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\n\nlet eval_Sub64 (dst src:nat64) : option (bool & nat64) =\n let diff = dst - src in Some (diff < 0, diff % pow2_64)\nval ins_Sub64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Sub64\n\nlet eval_Sbb64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let diff = dst - (src + (if old_carry then 1 else 0)) in\n Some (diff < 0, diff % pow2_64) // We specify cf, but underspecify everything else (via HavocFlags)\nval ins_Sbb64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_Sbb64\n\nlet eval_Mul64 (rax src:nat64) : option (nat64 & nat64) =\n Some (FStar.UInt.mul_div #64 rax src, FStar.UInt.mul_mod #64 rax src)\nval ins_Mul64 : instr_dep [out (one64Reg rRdx); inOut (one64Reg rRax)] [op64] HavocFlags eval_Mul64\n\nlet eval_Mulx64 (rdx src:nat64) : option (nat64 & nat64) =\n let hi = FStar.UInt.mul_div #64 rdx src in\n let lo = FStar.UInt.mul_mod #64 rdx src in\n if bmi2_enabled then Some (hi, lo) else None\nval ins_Mulx64 : instr_dep [out op64; out op64] [one64Reg rRdx; op64] PreserveFlags eval_Mulx64\n\nlet eval_IMul64 (dst src:nat64) : option nat64 =\n Some (FStar.UInt.mul_mod #64 dst src)\nval ins_IMul64 : instr_dep [inOut op64] [op64] HavocFlags eval_IMul64\n\nlet eval_And64 (dst src:nat64) : option nat64 = Some (iand dst src)\nval ins_And64 : instr_dep [inOut op64] [op64] HavocFlags eval_And64\n\nlet eval_Xor64 (dst src:nat64) : option (nat64 & (bool & bool)) =\n Some (Vale.Def.Types_s.ixor dst src, (false, false))\nval ins_Xor64 : instr_dep [inOut op64; out opFlagsCf; out opFlagsOf] [op64] HavocFlags eval_Xor64\n\nlet eval_Shr64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishr dst amt) else None\nval ins_Shr64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shr64\n\nlet eval_Shl64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishl dst amt) else None\nval ins_Shl64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shl64\n\nlet eval_Cpuid (rax rcx:nat64) : option (nat64 & (nat64 & (nat64 & nat64))) =\n Some (cpuid rRax rax rcx, (cpuid rRbx rax rcx, (cpuid rRcx rax rcx, cpuid rRdx rax rcx)))\nval ins_Cpuid :\n instr_dep\n [inOut (one64Reg rRax); out (one64Reg rRbx); inOut (one64Reg rRcx); out (one64Reg rRdx)]\n [] PreserveFlags eval_Cpuid\n\n// The XGETBV instruction requires that OSXSAVE (in CPUID) is enabled.\n// We underspecify XGETBV here to only support fetching XCR0, which\n// is supported on any processor supporting the XGETBV instruction\nlet eval_Xgetbv (rcx:nat64) : option (nat64 & nat64) =\n if osxsave_enabled && rcx = 0 then Some (xgetbv rRax rcx, xgetbv rRdx rcx) else None\nval ins_Xgetbv :\n instr_dep\n [out (one64Reg rRax); out (one64Reg rRdx)]\n [one64Reg rRcx] PreserveFlags eval_Xgetbv\n\nlet check_avx (#a:Type0) (x:option a) : option a =\n if avx_enabled then x else None\n\nlet check_sse2 (#a:Type0) (x:option a) : option a =\n if sse2_enabled then x else None\n\nlet check_ssse3 (#a:Type0) (x:option a) : option a =\n if ssse3_enabled then x else None\n\nlet check_sse4_1 (#a:Type0) (x:option a) : option a =\n if sse4_1_enabled then x else None\n\nlet eval_Movdqu (src:quad32) : option quad32 = check_sse2 (Some src)\nval ins_Movdqu : instr_dep [out opXmm] [opXmm] PreserveFlags eval_Movdqu\n\nlet eval_Pxor (dst src:quad32) : option quad32 = check_sse2 (Some (quad32_xor dst src))\nval ins_Pxor : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pxor\n\nlet eval_VPxor (src1 src2:quad32) : option quad32 = check_avx (Some (quad32_xor src1 src2))\nval ins_VPxor : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPxor\n\nlet eval_Pand (dst src:quad32) : option quad32 = check_sse2 (Some (four_map2 (fun di si -> iand di si) dst src))\nval ins_Pand : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pand\n\nlet eval_Paddd_raw (src1 src2:quad32) : option quad32 =\n Some (Mkfour\n ((src1.lo0 + src2.lo0) % pow2_32)\n ((src1.lo1 + src2.lo1) % pow2_32)\n ((src1.hi2 + src2.hi2) % pow2_32)\n ((src1.hi3 + src2.hi3) % pow2_32))\n\nlet eval_Paddd (src1 src2:quad32) : option quad32 = check_sse2 (eval_Paddd_raw src1 src2)\nval ins_Paddd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Paddd\n\nlet eval_VPaddd (src1 src2:quad32) : option quad32 = check_avx (eval_Paddd_raw src1 src2)\nval ins_VPaddd : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPaddd\n\nlet eval_Pslld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishl i amt) dst) else None)\nval ins_Pslld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Pslld amt)\n\nlet eval_Psrld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishr i amt) dst) else None)\nval ins_Psrld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrld amt)\n\nlet eval_Psrldq (amt:int) (dst:quad32) : option quad32 = check_sse2 (\n if 0 <= amt && amt < 16 then\n let src_bytes = le_quad32_to_bytes dst in\n let zero_pad = Seq.create amt 0 in\n let remaining_bytes = slice src_bytes amt (length src_bytes) in\n Some (le_bytes_to_quad32 (append zero_pad remaining_bytes))\n else None)\nval ins_Psrldq (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrldq amt)\n\nlet eval_Palignr_raw (amount:nat8) (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if amount = 4 then\n Some (Mkfour src2.lo1 src2.hi2 src2.hi3 src1.lo0)\n else if amount = 8 then\n Some (Mkfour src2.hi2 src2.hi3 src1.lo0 src1.lo1)\n else None\n\nlet eval_Palignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_ssse3 (eval_Palignr_raw amount src1 src2)\nval ins_Palignr (amount:nat8) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Palignr amount)\n\nlet eval_VPalignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Palignr_raw amount src1 src2)\nval ins_VPalignr (amount:nat8) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPalignr amount)\n\nlet eval_Shufpd_raw (permutation:int) (src1 src2:quad32) : option quad32 =\n if 0 <= permutation && permutation < 4 then\n Some (Mkfour\n (if permutation % 2 = 0 then src1.lo0 else src1.hi2)\n (if permutation % 2 = 0 then src1.lo1 else src1.hi3)\n (if (permutation / 2) % 2 = 0 then src2.lo0 else src2.hi2)\n (if (permutation / 2) % 2 = 0 then src2.lo1 else src2.hi3))\n else None\n\nlet eval_Shufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_sse2 (eval_Shufpd_raw permutation src1 src2)\nval ins_Shufpd (permutation:int) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Shufpd permutation)\n\nlet eval_VShufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Shufpd_raw permutation src1 src2)\nval ins_VShufpd (permutation:int) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VShufpd permutation)\n\nlet is_full_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x04050607 &&\n q.hi3 = 0x00010203\n\nlet is_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x00010203 &&\n q.lo1 = 0x04050607 &&\n q.hi2 = 0x08090A0B &&\n q.hi3 = 0x0C0D0E0F\n\nlet is_high_dup_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet is_lower_upper_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x04050607 &&\n q.lo1 = 0x00010203 &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet eval_Pshufb_raw (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if is_full_byte_reversal_mask src2 then\n Some (reverse_bytes_quad32 src1)\n else if is_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3))\n else if is_high_dup_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else if is_lower_upper_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else None\n\nlet eval_Pshufb (src1 src2:quad32) : option quad32 = check_ssse3 (eval_Pshufb_raw src1 src2)\nval ins_Pshufb : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pshufb\n\nlet eval_VPshufb (src1 src2:quad32) : option quad32 = check_avx (eval_Pshufb_raw src1 src2)\nval ins_VPshufb : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPshufb\n\nlet eval_Pshufd (permutation:nat8) (src:quad32) : option quad32 = check_sse2 (\n let bits:bits_of_byte = byte_to_twobits permutation in\n Some (Mkfour\n (select_word src bits.lo0)\n (select_word src bits.lo1)\n (select_word src bits.hi2)\n (select_word src bits.hi3)))\nval ins_Pshufd (permutation:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_Pshufd permutation)\n\nlet eval_Pcmpeqd (dst src:quad32) : option quad32 = check_sse2 (\n let eq_result (b:bool):nat32 = if b then 0xFFFFFFFF else 0 in\n Some (Mkfour\n (eq_result (src.lo0 = dst.lo0))\n (eq_result (src.lo1 = dst.lo1))\n (eq_result (src.hi2 = dst.hi2))\n (eq_result (src.hi3 = dst.hi3))))\nval ins_Pcmpeqd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pcmpeqd\n\nlet eval_Pextrq (index:nat8) (src:quad32) : option nat64 = check_sse4_1 (\n let src_two = four_to_two_two src in\n Some (two_to_nat 32 (two_select src_two (index % 2))))\nval ins_Pextrq (index:nat8) : instr_dep [out op64] [opXmm] PreserveFlags (eval_Pextrq index)\n\nlet eval_Pinsrd (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat32 dst (src % pow2_32) (index % 4)))\nval ins_Pinsrd (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrd index)\n\nlet eval_Pinsrq (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat64_def dst src (index % 2)))\nval ins_Pinsrq (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrq index)\n\nlet eval_Pslldq_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the two very special cases we need\n if count = 4 then Some (Mkfour 0 src.lo0 src.lo1 src.hi2)\n else if count = 8 then Some (Mkfour 0 0 src.lo0 src.lo1)\n else None\nlet eval_VPslldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Pslldq_raw count src)\nval ins_VPslldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPslldq count)\n\nlet eval_Psrldq_8_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the one very special case we need\n if count = 8 then Some (Mkfour src.hi2 src.hi3 0 0)\n else None\nlet eval_VPsrldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Psrldq_8_raw count src)\nval ins_VPsrldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPsrldq count)\n\nlet eval_Pclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n let Mkfour a0 a1 a2 a3 = src1 in\n let Mkfour b0 b1 b2 b3 = src2 in\n let f x0 x1 y0 y1 =\n let x = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo x0 x1) in\n let y = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo y0 y1) in\n Vale.Math.Poly2.Bits_s.to_quad32 (Vale.Math.Poly2_s.mul x y)\n in\n if pclmulqdq_enabled then\n match imm with\n | 0 -> Some (f a0 a1 b0 b1)\n | 1 -> Some (f a2 a3 b0 b1)\n | 16 -> Some (f a0 a1 b2 b3)\n | 17 -> Some (f a2 a3 b2 b3)\n | _ -> None\n else None\nval ins_Pclmulqdq (imm:int) : instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Pclmulqdq imm)\n\nlet eval_VPclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Pclmulqdq imm src1 src2)\nval ins_VPclmulqdq (imm:int) : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPclmulqdq imm)\n\nlet eval_AESNI_enc (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.mix_columns_LE (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1))) src2)\n else None\nval ins_AESNI_enc : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc\n\nlet eval_VAESNI_enc (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc src1 src2)\nval ins_VAESNI_enc : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc\n\nlet eval_AESNI_enc_last (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1)) src2)\n else None\nval ins_AESNI_enc_last : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc_last\n\nlet eval_VAESNI_enc_last (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc_last src1 src2)\nval ins_VAESNI_enc_last : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc_last\n\nlet eval_AESNI_dec (dst src:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.inv_mix_columns_LE (Vale.AES.AES_s.inv_sub_bytes (Vale.AES.AES_s.inv_shift_rows_LE dst))) src)\n else None\nval ins_AESNI_dec : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_dec\n\nlet eval_AESNI_dec_last (dst src:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.inv_sub_bytes (Vale.AES.AES_s.inv_shift_rows_LE dst)) src)\n else None\nval ins_AESNI_dec_last : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_dec_last\n\nlet eval_AESNI_imc (src:quad32) : option quad32 =\n if aesni_enabled then Some (Vale.AES.AES_s.inv_mix_columns_LE src) else None\nval ins_AESNI_imc : instr_dep [out opXmm] [opXmm] PreserveFlags eval_AESNI_imc\n\nlet eval_AESNI_keygen_assist (imm:nat8) (src:quad32) : option quad32 =\n if aesni_enabled then\n Some (Mkfour\n (Vale.AES.AES_s.sub_word src.lo1)\n (ixor (Vale.AES.AES_s.rot_word_LE (Vale.AES.AES_s.sub_word src.lo1)) imm)\n (Vale.AES.AES_s.sub_word src.hi3)\n (ixor (Vale.AES.AES_s.rot_word_LE (Vale.AES.AES_s.sub_word src.hi3)) imm))\n else None\nval ins_AESNI_keygen_assist (imm:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_AESNI_keygen_assist imm)\n\nlet eval_SHA256_rnds2 (src1 src2 wk:quad32) : option quad32 =\n if sha_enabled then Some (sha256_rnds2_spec src1 src2 wk) else None\nval ins_SHA256_rnds2 :\n instr_dep [inOut opXmm] [opXmm; oneXmm (OReg 0)] PreserveFlags eval_SHA256_rnds2\n\nlet eval_SHA256_msg1 (src1 src2:quad32) : option quad32 =\n if sha_enabled then Some (sha256_msg1_spec src1 src2) else None\nval ins_SHA256_msg1 : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_SHA256_msg1\n\nlet eval_SHA256_msg2 (src1 src2:quad32) : option quad32 =\n if sha_enabled then Some (sha256_msg2_spec src1 src2) else None\nval ins_SHA256_msg2 : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_SHA256_msg2\n\nlet eval_Ghost : option unit = Some ()\nval ins_Ghost : instr_dep [] [] PreserveFlags eval_Ghost\n\nlet eval_Comment : option unit = Some ()\nval ins_Comment (_:string) : instr_dep [] [] PreserveFlags eval_Comment\n\nlet eval_LargeComment : option unit = Some ()\nval ins_LargeComment (_:string) : instr_dep [] [] PreserveFlags eval_LargeComment\n\nPartial implementation:\nlet eval_Newline:option unit =\n \n"
},
{
"role": "assistant",
"content": "Some ()"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: src1: Vale.X64.Machine_s.quad32 -> src2: Vale.X64.Machine_s.quad32\n -> FStar.Pervasives.Native.option Vale.X64.Machine_s.quad32\nEffect: Prims.Tot\n\nContext:\nmodule Vale.X64.Instructions_s\nopen FStar.Mul\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Two_s\nopen Vale.Def.Words.Four_s\nopen Vale.Def.Types_s\nopen Vale.X64.Machine_s\nopen Vale.X64.Instruction_s\nopen Vale.X64.CPU_Features_s\nopen Vale.X64.CryptoInstructions_s\nopen FStar.Seq.Base\n\nlet eval_Mov64 (src:nat64) : option nat64 = Some src\nval ins_Mov64 : instr_dep [out op64] [op64] PreserveFlags eval_Mov64\n\nlet eval_MovBe64 (src:nat64) : option nat64 = if movbe_enabled then Some (reverse_bytes_nat64 src) else None\nval ins_MovBe64 : instr_dep [out op64] [op64] PreserveFlags eval_MovBe64\n\nlet eval_Bswap64 (dst:nat64) : option nat64 = Some (reverse_bytes_nat64 dst)\nval ins_Bswap64 : instr_dep [inOut op64] [] PreserveFlags eval_Bswap64\n\nlet eval_Cmovc64 (dst src:nat64) (carry:bool) : option nat64 = Some (if carry then src else dst)\nval ins_Cmovc64 : instr_dep [inOut op64] [op64; opFlagsCf] PreserveFlags eval_Cmovc64\n\nlet eval_Add64 (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src in Some (sum >= pow2_64, sum % pow2_64)\nval ins_Add64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Add64\n\nlet eval_AddLea64 (src1 src2:nat64) : option nat64 = Some ((src1 + src2) % pow2_64)\nval ins_AddLea64 :instr_dep [out op64] [op64; op64] PreserveFlags eval_AddLea64\n\nlet eval_AddCarry64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_carry then 1 else 0) in\n Some (sum >= pow2_64, sum % pow2_64)\nval ins_AddCarry64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_AddCarry64\n\nlet eval_Adcx64_Adox64 (old_flag:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_flag then 1 else 0) in\n if adx_enabled then Some (sum >= pow2_64, sum % pow2_64) else None\nval ins_Adcx64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\nval ins_Adox64 : instr_dep [inOut opFlagsOf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\n\nlet eval_Sub64 (dst src:nat64) : option (bool & nat64) =\n let diff = dst - src in Some (diff < 0, diff % pow2_64)\nval ins_Sub64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Sub64\n\nlet eval_Sbb64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let diff = dst - (src + (if old_carry then 1 else 0)) in\n Some (diff < 0, diff % pow2_64) // We specify cf, but underspecify everything else (via HavocFlags)\nval ins_Sbb64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_Sbb64\n\nlet eval_Mul64 (rax src:nat64) : option (nat64 & nat64) =\n Some (FStar.UInt.mul_div #64 rax src, FStar.UInt.mul_mod #64 rax src)\nval ins_Mul64 : instr_dep [out (one64Reg rRdx); inOut (one64Reg rRax)] [op64] HavocFlags eval_Mul64\n\nlet eval_Mulx64 (rdx src:nat64) : option (nat64 & nat64) =\n let hi = FStar.UInt.mul_div #64 rdx src in\n let lo = FStar.UInt.mul_mod #64 rdx src in\n if bmi2_enabled then Some (hi, lo) else None\nval ins_Mulx64 : instr_dep [out op64; out op64] [one64Reg rRdx; op64] PreserveFlags eval_Mulx64\n\nlet eval_IMul64 (dst src:nat64) : option nat64 =\n Some (FStar.UInt.mul_mod #64 dst src)\nval ins_IMul64 : instr_dep [inOut op64] [op64] HavocFlags eval_IMul64\n\nlet eval_And64 (dst src:nat64) : option nat64 = Some (iand dst src)\nval ins_And64 : instr_dep [inOut op64] [op64] HavocFlags eval_And64\n\nlet eval_Xor64 (dst src:nat64) : option (nat64 & (bool & bool)) =\n Some (Vale.Def.Types_s.ixor dst src, (false, false))\nval ins_Xor64 : instr_dep [inOut op64; out opFlagsCf; out opFlagsOf] [op64] HavocFlags eval_Xor64\n\nlet eval_Shr64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishr dst amt) else None\nval ins_Shr64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shr64\n\nlet eval_Shl64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishl dst amt) else None\nval ins_Shl64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shl64\n\nlet eval_Cpuid (rax rcx:nat64) : option (nat64 & (nat64 & (nat64 & nat64))) =\n Some (cpuid rRax rax rcx, (cpuid rRbx rax rcx, (cpuid rRcx rax rcx, cpuid rRdx rax rcx)))\nval ins_Cpuid :\n instr_dep\n [inOut (one64Reg rRax); out (one64Reg rRbx); inOut (one64Reg rRcx); out (one64Reg rRdx)]\n [] PreserveFlags eval_Cpuid\n\n// The XGETBV instruction requires that OSXSAVE (in CPUID) is enabled.\n// We underspecify XGETBV here to only support fetching XCR0, which\n// is supported on any processor supporting the XGETBV instruction\nlet eval_Xgetbv (rcx:nat64) : option (nat64 & nat64) =\n if osxsave_enabled && rcx = 0 then Some (xgetbv rRax rcx, xgetbv rRdx rcx) else None\nval ins_Xgetbv :\n instr_dep\n [out (one64Reg rRax); out (one64Reg rRdx)]\n [one64Reg rRcx] PreserveFlags eval_Xgetbv\n\nlet check_avx (#a:Type0) (x:option a) : option a =\n if avx_enabled then x else None\n\nlet check_sse2 (#a:Type0) (x:option a) : option a =\n if sse2_enabled then x else None\n\nlet check_ssse3 (#a:Type0) (x:option a) : option a =\n if ssse3_enabled then x else None\n\nlet check_sse4_1 (#a:Type0) (x:option a) : option a =\n if sse4_1_enabled then x else None\n\nlet eval_Movdqu (src:quad32) : option quad32 = check_sse2 (Some src)\nval ins_Movdqu : instr_dep [out opXmm] [opXmm] PreserveFlags eval_Movdqu\n\nlet eval_Pxor (dst src:quad32) : option quad32 = check_sse2 (Some (quad32_xor dst src))\nval ins_Pxor : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pxor\n\nlet eval_VPxor (src1 src2:quad32) : option quad32 = check_avx (Some (quad32_xor src1 src2))\nval ins_VPxor : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPxor\n\nlet eval_Pand (dst src:quad32) : option quad32 = check_sse2 (Some (four_map2 (fun di si -> iand di si) dst src))\nval ins_Pand : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pand\n\nlet eval_Paddd_raw (src1 src2:quad32) : option quad32 =\n Some (Mkfour\n ((src1.lo0 + src2.lo0) % pow2_32)\n ((src1.lo1 + src2.lo1) % pow2_32)\n ((src1.hi2 + src2.hi2) % pow2_32)\n ((src1.hi3 + src2.hi3) % pow2_32))\n\nlet eval_Paddd (src1 src2:quad32) : option quad32 = check_sse2 (eval_Paddd_raw src1 src2)\nval ins_Paddd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Paddd\n\nlet eval_VPaddd (src1 src2:quad32) : option quad32 = check_avx (eval_Paddd_raw src1 src2)\nval ins_VPaddd : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPaddd\n\nlet eval_Pslld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishl i amt) dst) else None)\nval ins_Pslld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Pslld amt)\n\nlet eval_Psrld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishr i amt) dst) else None)\nval ins_Psrld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrld amt)\n\nlet eval_Psrldq (amt:int) (dst:quad32) : option quad32 = check_sse2 (\n if 0 <= amt && amt < 16 then\n let src_bytes = le_quad32_to_bytes dst in\n let zero_pad = Seq.create amt 0 in\n let remaining_bytes = slice src_bytes amt (length src_bytes) in\n Some (le_bytes_to_quad32 (append zero_pad remaining_bytes))\n else None)\nval ins_Psrldq (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrldq amt)\n\nlet eval_Palignr_raw (amount:nat8) (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if amount = 4 then\n Some (Mkfour src2.lo1 src2.hi2 src2.hi3 src1.lo0)\n else if amount = 8 then\n Some (Mkfour src2.hi2 src2.hi3 src1.lo0 src1.lo1)\n else None\n\nlet eval_Palignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_ssse3 (eval_Palignr_raw amount src1 src2)\nval ins_Palignr (amount:nat8) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Palignr amount)\n\nlet eval_VPalignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Palignr_raw amount src1 src2)\nval ins_VPalignr (amount:nat8) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPalignr amount)\n\nlet eval_Shufpd_raw (permutation:int) (src1 src2:quad32) : option quad32 =\n if 0 <= permutation && permutation < 4 then\n Some (Mkfour\n (if permutation % 2 = 0 then src1.lo0 else src1.hi2)\n (if permutation % 2 = 0 then src1.lo1 else src1.hi3)\n (if (permutation / 2) % 2 = 0 then src2.lo0 else src2.hi2)\n (if (permutation / 2) % 2 = 0 then src2.lo1 else src2.hi3))\n else None\n\nlet eval_Shufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_sse2 (eval_Shufpd_raw permutation src1 src2)\nval ins_Shufpd (permutation:int) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Shufpd permutation)\n\nlet eval_VShufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Shufpd_raw permutation src1 src2)\nval ins_VShufpd (permutation:int) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VShufpd permutation)\n\nlet is_full_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x04050607 &&\n q.hi3 = 0x00010203\n\nlet is_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x00010203 &&\n q.lo1 = 0x04050607 &&\n q.hi2 = 0x08090A0B &&\n q.hi3 = 0x0C0D0E0F\n\nlet is_high_dup_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet is_lower_upper_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x04050607 &&\n q.lo1 = 0x00010203 &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet eval_Pshufb_raw (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if is_full_byte_reversal_mask src2 then\n Some (reverse_bytes_quad32 src1)\n else if is_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3))\n else if is_high_dup_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else if is_lower_upper_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else None\n\nlet eval_Pshufb (src1 src2:quad32) : option quad32 = check_ssse3 (eval_Pshufb_raw src1 src2)\nval ins_Pshufb : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pshufb\n\nlet eval_VPshufb (src1 src2:quad32) : option quad32 = check_avx (eval_Pshufb_raw src1 src2)\nval ins_VPshufb : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPshufb\n\nlet eval_Pshufd (permutation:nat8) (src:quad32) : option quad32 = check_sse2 (\n let bits:bits_of_byte = byte_to_twobits permutation in\n Some (Mkfour\n (select_word src bits.lo0)\n (select_word src bits.lo1)\n (select_word src bits.hi2)\n (select_word src bits.hi3)))\nval ins_Pshufd (permutation:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_Pshufd permutation)\n\nlet eval_Pcmpeqd (dst src:quad32) : option quad32 = check_sse2 (\n let eq_result (b:bool):nat32 = if b then 0xFFFFFFFF else 0 in\n Some (Mkfour\n (eq_result (src.lo0 = dst.lo0))\n (eq_result (src.lo1 = dst.lo1))\n (eq_result (src.hi2 = dst.hi2))\n (eq_result (src.hi3 = dst.hi3))))\nval ins_Pcmpeqd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pcmpeqd\n\nlet eval_Pextrq (index:nat8) (src:quad32) : option nat64 = check_sse4_1 (\n let src_two = four_to_two_two src in\n Some (two_to_nat 32 (two_select src_two (index % 2))))\nval ins_Pextrq (index:nat8) : instr_dep [out op64] [opXmm] PreserveFlags (eval_Pextrq index)\n\nlet eval_Pinsrd (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat32 dst (src % pow2_32) (index % 4)))\nval ins_Pinsrd (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrd index)\n\nlet eval_Pinsrq (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat64_def dst src (index % 2)))\nval ins_Pinsrq (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrq index)\n\nlet eval_Pslldq_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the two very special cases we need\n if count = 4 then Some (Mkfour 0 src.lo0 src.lo1 src.hi2)\n else if count = 8 then Some (Mkfour 0 0 src.lo0 src.lo1)\n else None\nlet eval_VPslldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Pslldq_raw count src)\nval ins_VPslldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPslldq count)\n\nlet eval_Psrldq_8_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the one very special case we need\n if count = 8 then Some (Mkfour src.hi2 src.hi3 0 0)\n else None\nlet eval_VPsrldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Psrldq_8_raw count src)\nval ins_VPsrldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPsrldq count)\n\nlet eval_Pclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n let Mkfour a0 a1 a2 a3 = src1 in\n let Mkfour b0 b1 b2 b3 = src2 in\n let f x0 x1 y0 y1 =\n let x = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo x0 x1) in\n let y = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo y0 y1) in\n Vale.Math.Poly2.Bits_s.to_quad32 (Vale.Math.Poly2_s.mul x y)\n in\n if pclmulqdq_enabled then\n match imm with\n | 0 -> Some (f a0 a1 b0 b1)\n | 1 -> Some (f a2 a3 b0 b1)\n | 16 -> Some (f a0 a1 b2 b3)\n | 17 -> Some (f a2 a3 b2 b3)\n | _ -> None\n else None\nval ins_Pclmulqdq (imm:int) : instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Pclmulqdq imm)\n\nlet eval_VPclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Pclmulqdq imm src1 src2)\nval ins_VPclmulqdq (imm:int) : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPclmulqdq imm)\n\nlet eval_AESNI_enc (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.mix_columns_LE (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1))) src2)\n else None\nval ins_AESNI_enc : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc\n\nlet eval_VAESNI_enc (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc src1 src2)\nval ins_VAESNI_enc : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc\n\nPartial implementation:\nlet eval_AESNI_enc_last (src1 src2: quad32) : option quad32 =\n \n"
},
{
"role": "assistant",
"content": "if aesni_enabled\nthen Some (quad32_xor (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1)) src2)\nelse None"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: FStar.Pervasives.Native.option Prims.unit\nEffect: Prims.Tot\n\nContext:\nmodule Vale.X64.Instructions_s\nopen FStar.Mul\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Two_s\nopen Vale.Def.Words.Four_s\nopen Vale.Def.Types_s\nopen Vale.X64.Machine_s\nopen Vale.X64.Instruction_s\nopen Vale.X64.CPU_Features_s\nopen Vale.X64.CryptoInstructions_s\nopen FStar.Seq.Base\n\nlet eval_Mov64 (src:nat64) : option nat64 = Some src\nval ins_Mov64 : instr_dep [out op64] [op64] PreserveFlags eval_Mov64\n\nlet eval_MovBe64 (src:nat64) : option nat64 = if movbe_enabled then Some (reverse_bytes_nat64 src) else None\nval ins_MovBe64 : instr_dep [out op64] [op64] PreserveFlags eval_MovBe64\n\nlet eval_Bswap64 (dst:nat64) : option nat64 = Some (reverse_bytes_nat64 dst)\nval ins_Bswap64 : instr_dep [inOut op64] [] PreserveFlags eval_Bswap64\n\nlet eval_Cmovc64 (dst src:nat64) (carry:bool) : option nat64 = Some (if carry then src else dst)\nval ins_Cmovc64 : instr_dep [inOut op64] [op64; opFlagsCf] PreserveFlags eval_Cmovc64\n\nlet eval_Add64 (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src in Some (sum >= pow2_64, sum % pow2_64)\nval ins_Add64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Add64\n\nlet eval_AddLea64 (src1 src2:nat64) : option nat64 = Some ((src1 + src2) % pow2_64)\nval ins_AddLea64 :instr_dep [out op64] [op64; op64] PreserveFlags eval_AddLea64\n\nlet eval_AddCarry64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_carry then 1 else 0) in\n Some (sum >= pow2_64, sum % pow2_64)\nval ins_AddCarry64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_AddCarry64\n\nlet eval_Adcx64_Adox64 (old_flag:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_flag then 1 else 0) in\n if adx_enabled then Some (sum >= pow2_64, sum % pow2_64) else None\nval ins_Adcx64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\nval ins_Adox64 : instr_dep [inOut opFlagsOf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\n\nlet eval_Sub64 (dst src:nat64) : option (bool & nat64) =\n let diff = dst - src in Some (diff < 0, diff % pow2_64)\nval ins_Sub64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Sub64\n\nlet eval_Sbb64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let diff = dst - (src + (if old_carry then 1 else 0)) in\n Some (diff < 0, diff % pow2_64) // We specify cf, but underspecify everything else (via HavocFlags)\nval ins_Sbb64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_Sbb64\n\nlet eval_Mul64 (rax src:nat64) : option (nat64 & nat64) =\n Some (FStar.UInt.mul_div #64 rax src, FStar.UInt.mul_mod #64 rax src)\nval ins_Mul64 : instr_dep [out (one64Reg rRdx); inOut (one64Reg rRax)] [op64] HavocFlags eval_Mul64\n\nlet eval_Mulx64 (rdx src:nat64) : option (nat64 & nat64) =\n let hi = FStar.UInt.mul_div #64 rdx src in\n let lo = FStar.UInt.mul_mod #64 rdx src in\n if bmi2_enabled then Some (hi, lo) else None\nval ins_Mulx64 : instr_dep [out op64; out op64] [one64Reg rRdx; op64] PreserveFlags eval_Mulx64\n\nlet eval_IMul64 (dst src:nat64) : option nat64 =\n Some (FStar.UInt.mul_mod #64 dst src)\nval ins_IMul64 : instr_dep [inOut op64] [op64] HavocFlags eval_IMul64\n\nlet eval_And64 (dst src:nat64) : option nat64 = Some (iand dst src)\nval ins_And64 : instr_dep [inOut op64] [op64] HavocFlags eval_And64\n\nlet eval_Xor64 (dst src:nat64) : option (nat64 & (bool & bool)) =\n Some (Vale.Def.Types_s.ixor dst src, (false, false))\nval ins_Xor64 : instr_dep [inOut op64; out opFlagsCf; out opFlagsOf] [op64] HavocFlags eval_Xor64\n\nlet eval_Shr64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishr dst amt) else None\nval ins_Shr64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shr64\n\nlet eval_Shl64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishl dst amt) else None\nval ins_Shl64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shl64\n\nlet eval_Cpuid (rax rcx:nat64) : option (nat64 & (nat64 & (nat64 & nat64))) =\n Some (cpuid rRax rax rcx, (cpuid rRbx rax rcx, (cpuid rRcx rax rcx, cpuid rRdx rax rcx)))\nval ins_Cpuid :\n instr_dep\n [inOut (one64Reg rRax); out (one64Reg rRbx); inOut (one64Reg rRcx); out (one64Reg rRdx)]\n [] PreserveFlags eval_Cpuid\n\n// The XGETBV instruction requires that OSXSAVE (in CPUID) is enabled.\n// We underspecify XGETBV here to only support fetching XCR0, which\n// is supported on any processor supporting the XGETBV instruction\nlet eval_Xgetbv (rcx:nat64) : option (nat64 & nat64) =\n if osxsave_enabled && rcx = 0 then Some (xgetbv rRax rcx, xgetbv rRdx rcx) else None\nval ins_Xgetbv :\n instr_dep\n [out (one64Reg rRax); out (one64Reg rRdx)]\n [one64Reg rRcx] PreserveFlags eval_Xgetbv\n\nlet check_avx (#a:Type0) (x:option a) : option a =\n if avx_enabled then x else None\n\nlet check_sse2 (#a:Type0) (x:option a) : option a =\n if sse2_enabled then x else None\n\nlet check_ssse3 (#a:Type0) (x:option a) : option a =\n if ssse3_enabled then x else None\n\nlet check_sse4_1 (#a:Type0) (x:option a) : option a =\n if sse4_1_enabled then x else None\n\nlet eval_Movdqu (src:quad32) : option quad32 = check_sse2 (Some src)\nval ins_Movdqu : instr_dep [out opXmm] [opXmm] PreserveFlags eval_Movdqu\n\nlet eval_Pxor (dst src:quad32) : option quad32 = check_sse2 (Some (quad32_xor dst src))\nval ins_Pxor : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pxor\n\nlet eval_VPxor (src1 src2:quad32) : option quad32 = check_avx (Some (quad32_xor src1 src2))\nval ins_VPxor : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPxor\n\nlet eval_Pand (dst src:quad32) : option quad32 = check_sse2 (Some (four_map2 (fun di si -> iand di si) dst src))\nval ins_Pand : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pand\n\nlet eval_Paddd_raw (src1 src2:quad32) : option quad32 =\n Some (Mkfour\n ((src1.lo0 + src2.lo0) % pow2_32)\n ((src1.lo1 + src2.lo1) % pow2_32)\n ((src1.hi2 + src2.hi2) % pow2_32)\n ((src1.hi3 + src2.hi3) % pow2_32))\n\nlet eval_Paddd (src1 src2:quad32) : option quad32 = check_sse2 (eval_Paddd_raw src1 src2)\nval ins_Paddd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Paddd\n\nlet eval_VPaddd (src1 src2:quad32) : option quad32 = check_avx (eval_Paddd_raw src1 src2)\nval ins_VPaddd : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPaddd\n\nlet eval_Pslld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishl i amt) dst) else None)\nval ins_Pslld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Pslld amt)\n\nlet eval_Psrld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishr i amt) dst) else None)\nval ins_Psrld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrld amt)\n\nlet eval_Psrldq (amt:int) (dst:quad32) : option quad32 = check_sse2 (\n if 0 <= amt && amt < 16 then\n let src_bytes = le_quad32_to_bytes dst in\n let zero_pad = Seq.create amt 0 in\n let remaining_bytes = slice src_bytes amt (length src_bytes) in\n Some (le_bytes_to_quad32 (append zero_pad remaining_bytes))\n else None)\nval ins_Psrldq (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrldq amt)\n\nlet eval_Palignr_raw (amount:nat8) (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if amount = 4 then\n Some (Mkfour src2.lo1 src2.hi2 src2.hi3 src1.lo0)\n else if amount = 8 then\n Some (Mkfour src2.hi2 src2.hi3 src1.lo0 src1.lo1)\n else None\n\nlet eval_Palignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_ssse3 (eval_Palignr_raw amount src1 src2)\nval ins_Palignr (amount:nat8) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Palignr amount)\n\nlet eval_VPalignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Palignr_raw amount src1 src2)\nval ins_VPalignr (amount:nat8) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPalignr amount)\n\nlet eval_Shufpd_raw (permutation:int) (src1 src2:quad32) : option quad32 =\n if 0 <= permutation && permutation < 4 then\n Some (Mkfour\n (if permutation % 2 = 0 then src1.lo0 else src1.hi2)\n (if permutation % 2 = 0 then src1.lo1 else src1.hi3)\n (if (permutation / 2) % 2 = 0 then src2.lo0 else src2.hi2)\n (if (permutation / 2) % 2 = 0 then src2.lo1 else src2.hi3))\n else None\n\nlet eval_Shufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_sse2 (eval_Shufpd_raw permutation src1 src2)\nval ins_Shufpd (permutation:int) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Shufpd permutation)\n\nlet eval_VShufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Shufpd_raw permutation src1 src2)\nval ins_VShufpd (permutation:int) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VShufpd permutation)\n\nlet is_full_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x04050607 &&\n q.hi3 = 0x00010203\n\nlet is_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x00010203 &&\n q.lo1 = 0x04050607 &&\n q.hi2 = 0x08090A0B &&\n q.hi3 = 0x0C0D0E0F\n\nlet is_high_dup_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet is_lower_upper_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x04050607 &&\n q.lo1 = 0x00010203 &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet eval_Pshufb_raw (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if is_full_byte_reversal_mask src2 then\n Some (reverse_bytes_quad32 src1)\n else if is_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3))\n else if is_high_dup_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else if is_lower_upper_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else None\n\nlet eval_Pshufb (src1 src2:quad32) : option quad32 = check_ssse3 (eval_Pshufb_raw src1 src2)\nval ins_Pshufb : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pshufb\n\nlet eval_VPshufb (src1 src2:quad32) : option quad32 = check_avx (eval_Pshufb_raw src1 src2)\nval ins_VPshufb : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPshufb\n\nlet eval_Pshufd (permutation:nat8) (src:quad32) : option quad32 = check_sse2 (\n let bits:bits_of_byte = byte_to_twobits permutation in\n Some (Mkfour\n (select_word src bits.lo0)\n (select_word src bits.lo1)\n (select_word src bits.hi2)\n (select_word src bits.hi3)))\nval ins_Pshufd (permutation:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_Pshufd permutation)\n\nlet eval_Pcmpeqd (dst src:quad32) : option quad32 = check_sse2 (\n let eq_result (b:bool):nat32 = if b then 0xFFFFFFFF else 0 in\n Some (Mkfour\n (eq_result (src.lo0 = dst.lo0))\n (eq_result (src.lo1 = dst.lo1))\n (eq_result (src.hi2 = dst.hi2))\n (eq_result (src.hi3 = dst.hi3))))\nval ins_Pcmpeqd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pcmpeqd\n\nlet eval_Pextrq (index:nat8) (src:quad32) : option nat64 = check_sse4_1 (\n let src_two = four_to_two_two src in\n Some (two_to_nat 32 (two_select src_two (index % 2))))\nval ins_Pextrq (index:nat8) : instr_dep [out op64] [opXmm] PreserveFlags (eval_Pextrq index)\n\nlet eval_Pinsrd (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat32 dst (src % pow2_32) (index % 4)))\nval ins_Pinsrd (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrd index)\n\nlet eval_Pinsrq (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat64_def dst src (index % 2)))\nval ins_Pinsrq (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrq index)\n\nlet eval_Pslldq_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the two very special cases we need\n if count = 4 then Some (Mkfour 0 src.lo0 src.lo1 src.hi2)\n else if count = 8 then Some (Mkfour 0 0 src.lo0 src.lo1)\n else None\nlet eval_VPslldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Pslldq_raw count src)\nval ins_VPslldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPslldq count)\n\nlet eval_Psrldq_8_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the one very special case we need\n if count = 8 then Some (Mkfour src.hi2 src.hi3 0 0)\n else None\nlet eval_VPsrldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Psrldq_8_raw count src)\nval ins_VPsrldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPsrldq count)\n\nlet eval_Pclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n let Mkfour a0 a1 a2 a3 = src1 in\n let Mkfour b0 b1 b2 b3 = src2 in\n let f x0 x1 y0 y1 =\n let x = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo x0 x1) in\n let y = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo y0 y1) in\n Vale.Math.Poly2.Bits_s.to_quad32 (Vale.Math.Poly2_s.mul x y)\n in\n if pclmulqdq_enabled then\n match imm with\n | 0 -> Some (f a0 a1 b0 b1)\n | 1 -> Some (f a2 a3 b0 b1)\n | 16 -> Some (f a0 a1 b2 b3)\n | 17 -> Some (f a2 a3 b2 b3)\n | _ -> None\n else None\nval ins_Pclmulqdq (imm:int) : instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Pclmulqdq imm)\n\nlet eval_VPclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Pclmulqdq imm src1 src2)\nval ins_VPclmulqdq (imm:int) : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPclmulqdq imm)\n\nlet eval_AESNI_enc (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.mix_columns_LE (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1))) src2)\n else None\nval ins_AESNI_enc : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc\n\nlet eval_VAESNI_enc (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc src1 src2)\nval ins_VAESNI_enc : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc\n\nlet eval_AESNI_enc_last (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1)) src2)\n else None\nval ins_AESNI_enc_last : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc_last\n\nlet eval_VAESNI_enc_last (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc_last src1 src2)\nval ins_VAESNI_enc_last : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc_last\n\nlet eval_AESNI_dec (dst src:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.inv_mix_columns_LE (Vale.AES.AES_s.inv_sub_bytes (Vale.AES.AES_s.inv_shift_rows_LE dst))) src)\n else None\nval ins_AESNI_dec : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_dec\n\nlet eval_AESNI_dec_last (dst src:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.inv_sub_bytes (Vale.AES.AES_s.inv_shift_rows_LE dst)) src)\n else None\nval ins_AESNI_dec_last : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_dec_last\n\nlet eval_AESNI_imc (src:quad32) : option quad32 =\n if aesni_enabled then Some (Vale.AES.AES_s.inv_mix_columns_LE src) else None\nval ins_AESNI_imc : instr_dep [out opXmm] [opXmm] PreserveFlags eval_AESNI_imc\n\nlet eval_AESNI_keygen_assist (imm:nat8) (src:quad32) : option quad32 =\n if aesni_enabled then\n Some (Mkfour\n (Vale.AES.AES_s.sub_word src.lo1)\n (ixor (Vale.AES.AES_s.rot_word_LE (Vale.AES.AES_s.sub_word src.lo1)) imm)\n (Vale.AES.AES_s.sub_word src.hi3)\n (ixor (Vale.AES.AES_s.rot_word_LE (Vale.AES.AES_s.sub_word src.hi3)) imm))\n else None\nval ins_AESNI_keygen_assist (imm:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_AESNI_keygen_assist imm)\n\nlet eval_SHA256_rnds2 (src1 src2 wk:quad32) : option quad32 =\n if sha_enabled then Some (sha256_rnds2_spec src1 src2 wk) else None\nval ins_SHA256_rnds2 :\n instr_dep [inOut opXmm] [opXmm; oneXmm (OReg 0)] PreserveFlags eval_SHA256_rnds2\n\nlet eval_SHA256_msg1 (src1 src2:quad32) : option quad32 =\n if sha_enabled then Some (sha256_msg1_spec src1 src2) else None\nval ins_SHA256_msg1 : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_SHA256_msg1\n\nlet eval_SHA256_msg2 (src1 src2:quad32) : option quad32 =\n if sha_enabled then Some (sha256_msg2_spec src1 src2) else None\nval ins_SHA256_msg2 : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_SHA256_msg2\n\nlet eval_Ghost : option unit = Some ()\nval ins_Ghost : instr_dep [] [] PreserveFlags eval_Ghost\n\nlet eval_Comment : option unit = Some ()\nval ins_Comment (_:string) : instr_dep [] [] PreserveFlags eval_Comment\n\nlet eval_LargeComment : option unit = Some ()\nval ins_LargeComment (_:string) : instr_dep [] [] PreserveFlags eval_LargeComment\n\nlet eval_Newline : option unit = Some ()\nval ins_Newline : instr_dep [] [] PreserveFlags eval_Newline\n\nPartial implementation:\nlet eval_Space:option unit =\n \n"
},
{
"role": "assistant",
"content": "Some ()"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: Prims.int\nEffect: Prims.Tot\n\nContext:\nmodule Spec.Box\n\nopen FStar.Mul\nopen Lib.IntTypes\nopen Lib.Sequence\nopen Lib.ByteSequence\n\nopen Spec.SecretBox\n\n#set-options \"--z3rlimit 20 --max_fuel 0 --max_ifuel 0\"\n\nPartial implementation:\nlet size_publickey =\n \n"
},
{
"role": "assistant",
"content": "32"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: Prims.int\nEffect: Prims.Tot\n\nContext:\nmodule Spec.Box\n\nopen FStar.Mul\nopen Lib.IntTypes\nopen Lib.Sequence\nopen Lib.ByteSequence\n\nopen Spec.SecretBox\n\n#set-options \"--z3rlimit 20 --max_fuel 0 --max_ifuel 0\"\n\n(* Constants *)\n\nPartial implementation:\nlet size_secretkey =\n \n"
},
{
"role": "assistant",
"content": "32"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: amt: Prims.int -> dst: Vale.X64.Machine_s.quad32\n -> FStar.Pervasives.Native.option Vale.X64.Machine_s.quad32\nEffect: Prims.Tot\n\nContext:\nmodule Vale.X64.Instructions_s\nopen FStar.Mul\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Two_s\nopen Vale.Def.Words.Four_s\nopen Vale.Def.Types_s\nopen Vale.X64.Machine_s\nopen Vale.X64.Instruction_s\nopen Vale.X64.CPU_Features_s\nopen Vale.X64.CryptoInstructions_s\nopen FStar.Seq.Base\n\nlet eval_Mov64 (src:nat64) : option nat64 = Some src\nval ins_Mov64 : instr_dep [out op64] [op64] PreserveFlags eval_Mov64\n\nlet eval_MovBe64 (src:nat64) : option nat64 = if movbe_enabled then Some (reverse_bytes_nat64 src) else None\nval ins_MovBe64 : instr_dep [out op64] [op64] PreserveFlags eval_MovBe64\n\nlet eval_Bswap64 (dst:nat64) : option nat64 = Some (reverse_bytes_nat64 dst)\nval ins_Bswap64 : instr_dep [inOut op64] [] PreserveFlags eval_Bswap64\n\nlet eval_Cmovc64 (dst src:nat64) (carry:bool) : option nat64 = Some (if carry then src else dst)\nval ins_Cmovc64 : instr_dep [inOut op64] [op64; opFlagsCf] PreserveFlags eval_Cmovc64\n\nlet eval_Add64 (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src in Some (sum >= pow2_64, sum % pow2_64)\nval ins_Add64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Add64\n\nlet eval_AddLea64 (src1 src2:nat64) : option nat64 = Some ((src1 + src2) % pow2_64)\nval ins_AddLea64 :instr_dep [out op64] [op64; op64] PreserveFlags eval_AddLea64\n\nlet eval_AddCarry64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_carry then 1 else 0) in\n Some (sum >= pow2_64, sum % pow2_64)\nval ins_AddCarry64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_AddCarry64\n\nlet eval_Adcx64_Adox64 (old_flag:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_flag then 1 else 0) in\n if adx_enabled then Some (sum >= pow2_64, sum % pow2_64) else None\nval ins_Adcx64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\nval ins_Adox64 : instr_dep [inOut opFlagsOf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\n\nlet eval_Sub64 (dst src:nat64) : option (bool & nat64) =\n let diff = dst - src in Some (diff < 0, diff % pow2_64)\nval ins_Sub64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Sub64\n\nlet eval_Sbb64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let diff = dst - (src + (if old_carry then 1 else 0)) in\n Some (diff < 0, diff % pow2_64) // We specify cf, but underspecify everything else (via HavocFlags)\nval ins_Sbb64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_Sbb64\n\nlet eval_Mul64 (rax src:nat64) : option (nat64 & nat64) =\n Some (FStar.UInt.mul_div #64 rax src, FStar.UInt.mul_mod #64 rax src)\nval ins_Mul64 : instr_dep [out (one64Reg rRdx); inOut (one64Reg rRax)] [op64] HavocFlags eval_Mul64\n\nlet eval_Mulx64 (rdx src:nat64) : option (nat64 & nat64) =\n let hi = FStar.UInt.mul_div #64 rdx src in\n let lo = FStar.UInt.mul_mod #64 rdx src in\n if bmi2_enabled then Some (hi, lo) else None\nval ins_Mulx64 : instr_dep [out op64; out op64] [one64Reg rRdx; op64] PreserveFlags eval_Mulx64\n\nlet eval_IMul64 (dst src:nat64) : option nat64 =\n Some (FStar.UInt.mul_mod #64 dst src)\nval ins_IMul64 : instr_dep [inOut op64] [op64] HavocFlags eval_IMul64\n\nlet eval_And64 (dst src:nat64) : option nat64 = Some (iand dst src)\nval ins_And64 : instr_dep [inOut op64] [op64] HavocFlags eval_And64\n\nlet eval_Xor64 (dst src:nat64) : option (nat64 & (bool & bool)) =\n Some (Vale.Def.Types_s.ixor dst src, (false, false))\nval ins_Xor64 : instr_dep [inOut op64; out opFlagsCf; out opFlagsOf] [op64] HavocFlags eval_Xor64\n\nlet eval_Shr64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishr dst amt) else None\nval ins_Shr64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shr64\n\nlet eval_Shl64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishl dst amt) else None\nval ins_Shl64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shl64\n\nlet eval_Cpuid (rax rcx:nat64) : option (nat64 & (nat64 & (nat64 & nat64))) =\n Some (cpuid rRax rax rcx, (cpuid rRbx rax rcx, (cpuid rRcx rax rcx, cpuid rRdx rax rcx)))\nval ins_Cpuid :\n instr_dep\n [inOut (one64Reg rRax); out (one64Reg rRbx); inOut (one64Reg rRcx); out (one64Reg rRdx)]\n [] PreserveFlags eval_Cpuid\n\n// The XGETBV instruction requires that OSXSAVE (in CPUID) is enabled.\n// We underspecify XGETBV here to only support fetching XCR0, which\n// is supported on any processor supporting the XGETBV instruction\nlet eval_Xgetbv (rcx:nat64) : option (nat64 & nat64) =\n if osxsave_enabled && rcx = 0 then Some (xgetbv rRax rcx, xgetbv rRdx rcx) else None\nval ins_Xgetbv :\n instr_dep\n [out (one64Reg rRax); out (one64Reg rRdx)]\n [one64Reg rRcx] PreserveFlags eval_Xgetbv\n\nlet check_avx (#a:Type0) (x:option a) : option a =\n if avx_enabled then x else None\n\nlet check_sse2 (#a:Type0) (x:option a) : option a =\n if sse2_enabled then x else None\n\nlet check_ssse3 (#a:Type0) (x:option a) : option a =\n if ssse3_enabled then x else None\n\nlet check_sse4_1 (#a:Type0) (x:option a) : option a =\n if sse4_1_enabled then x else None\n\nlet eval_Movdqu (src:quad32) : option quad32 = check_sse2 (Some src)\nval ins_Movdqu : instr_dep [out opXmm] [opXmm] PreserveFlags eval_Movdqu\n\nlet eval_Pxor (dst src:quad32) : option quad32 = check_sse2 (Some (quad32_xor dst src))\nval ins_Pxor : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pxor\n\nlet eval_VPxor (src1 src2:quad32) : option quad32 = check_avx (Some (quad32_xor src1 src2))\nval ins_VPxor : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPxor\n\nlet eval_Pand (dst src:quad32) : option quad32 = check_sse2 (Some (four_map2 (fun di si -> iand di si) dst src))\nval ins_Pand : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pand\n\nlet eval_Paddd_raw (src1 src2:quad32) : option quad32 =\n Some (Mkfour\n ((src1.lo0 + src2.lo0) % pow2_32)\n ((src1.lo1 + src2.lo1) % pow2_32)\n ((src1.hi2 + src2.hi2) % pow2_32)\n ((src1.hi3 + src2.hi3) % pow2_32))\n\nlet eval_Paddd (src1 src2:quad32) : option quad32 = check_sse2 (eval_Paddd_raw src1 src2)\nval ins_Paddd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Paddd\n\nlet eval_VPaddd (src1 src2:quad32) : option quad32 = check_avx (eval_Paddd_raw src1 src2)\nval ins_VPaddd : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPaddd\n\nlet eval_Pslld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishl i amt) dst) else None)\nval ins_Pslld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Pslld amt)\n\nlet eval_Psrld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishr i amt) dst) else None)\nval ins_Psrld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrld amt)\n\nPartial implementation:\nlet eval_Psrldq (amt: int) (dst: quad32) : option quad32 =\n \n"
},
{
"role": "assistant",
"content": "check_sse2 (if 0 <= amt && amt < 16\n then\n let src_bytes = le_quad32_to_bytes dst in\n let zero_pad = Seq.create amt 0 in\n let remaining_bytes = slice src_bytes amt (length src_bytes) in\n Some (le_bytes_to_quad32 (append zero_pad remaining_bytes))\n else None)"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n alg: Vale.AES.AES_common_s.algorithm ->\n offset: Prims.int ->\n auth_b: Vale.X64.Memory.buffer128 ->\n abytes_b: Vale.X64.Memory.buffer128 ->\n in128x6_b: Vale.X64.Memory.buffer128 ->\n out128x6_b: Vale.X64.Memory.buffer128 ->\n in128_b: Vale.X64.Memory.buffer128 ->\n out128_b: Vale.X64.Memory.buffer128 ->\n inout_b: Vale.X64.Memory.buffer128 ->\n iv_b: Vale.X64.Memory.buffer128 ->\n scratch_b: Vale.X64.Memory.buffer128 ->\n key: FStar.Seq.Base.seq Vale.X64.Memory.nat32 ->\n round_keys: FStar.Seq.Base.seq Vale.X64.Decls.quad32 ->\n keys_b: Vale.X64.Memory.buffer128 ->\n hkeys_b: Vale.X64.Memory.buffer128\n -> Vale.X64.QuickCode.va_quickCode Prims.unit\n (Vale.AES.X64.GCMdecryptOpt.va_code_Gcm_blocks alg offset)\nEffect: Prims.Tot\n\nContext:\nmodule Vale.AES.X64.GCMdecryptOpt\nopen Vale.Def.Prop_s\nopen Vale.Def.Opaque_s\nopen FStar.Seq\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Seq_s\nopen Vale.Def.Types_s\nopen Vale.Arch.Types\nopen Vale.Arch.HeapImpl\nopen Vale.AES.AES_s\nopen Vale.AES.GCTR_s\nopen Vale.AES.GCTR\nopen Vale.AES.GCM\nopen Vale.AES.GHash_s\nopen Vale.AES.GHash\nopen Vale.AES.GCM_s\nopen Vale.AES.X64.AES\nopen Vale.AES.GF128_s\nopen Vale.AES.GF128\nopen Vale.Poly1305.Math\nopen Vale.AES.GCM_helpers\nopen Vale.AES.X64.GHash\nopen Vale.AES.X64.GCTR\nopen Vale.X64.Machine_s\nopen Vale.X64.Memory\nopen Vale.X64.Stack_i\nopen Vale.X64.State\nopen Vale.X64.Decls\nopen Vale.X64.InsBasic\nopen Vale.X64.InsMem\nopen Vale.X64.InsVector\nopen Vale.X64.InsStack\nopen Vale.X64.InsAes\nopen Vale.X64.QuickCode\nopen Vale.X64.QuickCodes\nopen Vale.AES.X64.GF128_Mul\nopen Vale.X64.Stack\nopen Vale.X64.CPU_Features_s\nopen Vale.Math.Poly2.Bits_s\nopen Vale.AES.X64.AESopt\nopen Vale.AES.X64.AESGCM\nopen Vale.AES.X64.AESopt2\nopen Vale.Lib.Meta\nopen Vale.AES.X64.GCMencryptOpt\nopen Vale.AES.OptPublic\nopen Vale.Lib.Basic\n#reset-options \"--z3rlimit 20 --max_ifuel 0\"\n//-- Gcm_extra_bytes\n\nval va_code_Gcm_extra_bytes : alg:algorithm -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_extra_bytes alg =\n (va_Block (va_CCons (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0)\n (va_op_reg_opr64_reg64 rRax) 0 Secret) (va_CCons (va_code_Mov128 (va_op_xmm_xmm 10)\n (va_op_xmm_xmm 0)) (va_CCons (va_code_Ghash_extra_bytes ()) (va_CCons (va_code_Mov128\n (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_CCons (va_code_Pshufb (va_op_xmm_xmm 0)\n (va_op_xmm_xmm 9)) (va_CCons (va_code_AESEncryptBlock alg) (va_CCons (va_code_Pxor\n (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (va_CCons (va_code_Store128_buffer\n (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax) (va_op_xmm_xmm 10) 0 Secret)\n (va_CNil ()))))))))))\n\nval va_codegen_success_Gcm_extra_bytes : alg:algorithm -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_extra_bytes alg =\n (va_pbool_and (va_codegen_success_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0)\n (va_op_reg_opr64_reg64 rRax) 0 Secret) (va_pbool_and (va_codegen_success_Mov128 (va_op_xmm_xmm\n 10) (va_op_xmm_xmm 0)) (va_pbool_and (va_codegen_success_Ghash_extra_bytes ()) (va_pbool_and\n (va_codegen_success_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_pbool_and\n (va_codegen_success_Pshufb (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (va_pbool_and\n (va_codegen_success_AESEncryptBlock alg) (va_pbool_and (va_codegen_success_Pxor (va_op_xmm_xmm\n 10) (va_op_xmm_xmm 0)) (va_pbool_and (va_codegen_success_Store128_buffer\n (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax) (va_op_xmm_xmm 10) 0 Secret)\n (va_ttrue ())))))))))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_extra_bytes (va_mods:va_mods_t) (alg:algorithm) (inout_b:buffer128) (key:(seq\n nat32)) (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (total_bytes:nat)\n (old_hash:quad32) (completed_quads:(seq quad32)) (h_LE:quad32) : (va_quickCode unit\n (va_code_Gcm_extra_bytes alg)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in let (len:(va_int_range\n 1 1)) = 1 in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 188 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64\n rRax) 0 Secret inout_b 0) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 189 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov128 (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (fun (va_s:va_state) _ -> let\n (hash_input:quad32) = va_get_xmm 0 va_s in va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 193 column 22 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Ghash_extra_bytes hkeys_b total_bytes old_hash h_LE completed_quads) (fun\n (va_s:va_state) _ -> va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 194 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (FStar.Seq.Base.equal #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s)\n inout_b) (FStar.Seq.Base.create #quad32 1 hash_input)) (let (snap:(FStar.Seq.Base.seq\n Vale.X64.Decls.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s) inout_b in va_QSeq\n va_range1\n \"***** PRECONDITION NOT MET AT line 198 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 199 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Pshufb (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (fun (va_s:va_state) _ -> va_QBind\n va_range1\n \"***** PRECONDITION NOT MET AT line 200 column 20 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_AESEncryptBlock alg (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 11 va_s)) key\n round_keys keys_b) (fun (va_s:va_state) _ -> va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 201 column 26 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.AES_s.aes_encrypt_LE_reveal ()) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 204 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Pxor (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 205 column 20 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Store128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax)\n (va_op_xmm_xmm 10) 0 Secret inout_b 0) (fun (va_s:va_state) _ -> va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 207 column 24 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.gctr_partial_reveal ()) (va_QEmpty (()))))))))))))))\n\n\nval va_lemma_Gcm_extra_bytes : va_b0:va_code -> va_s0:va_state -> alg:algorithm ->\n inout_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128 -> total_bytes:nat -> old_hash:quad32 -> completed_quads:(seq quad32) ->\n h_LE:quad32\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_extra_bytes alg) va_s0 /\\ va_get_ok va_s0 /\\ (let\n (len:(va_int_range 1 1)) = 1 in sse_enabled /\\ Vale.X64.Decls.buffers_disjoint128 keys_b\n inout_b /\\ Vale.X64.Decls.buffers_disjoint128 hkeys_b inout_b /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) (va_get_reg64 rRax va_s0) inout_b\n len (va_get_mem_layout va_s0) Secret /\\ len == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 inout_b /\\ va_get_xmm 9 va_s0 == Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\ aes_reqs alg key round_keys\n keys_b (va_get_reg64 rR8 va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\\n pclmulqdq_enabled /\\ Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0\n va_s0) hkeys_b) (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret /\\ va_get_xmm 8 va_s0 == Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.AES.GHash.ghash_incremental0 h_LE old_hash completed_quads) /\\ FStar.Seq.Base.length\n #quad32 completed_quads == total_bytes `op_Division` 16 /\\ total_bytes < 16 `op_Multiply`\n FStar.Seq.Base.length #quad32 completed_quads + 16 /\\ va_get_reg64 rR10 va_s0 == total_bytes\n `op_Modulus` 16 /\\ total_bytes `op_Modulus` 16 =!= 0 /\\ (0 < total_bytes /\\ total_bytes < 16\n `op_Multiply` Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes) /\\ 16 `op_Multiply`\n (Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes - 1) < total_bytes)))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (let (len:(va_int_range 1 1)) = 1 in Vale.X64.Decls.modifies_buffer128 inout_b\n (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5 va_sM) /\\ Vale.AES.GCTR.gctr_partial alg len\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b) key (va_get_xmm 11 va_s0) /\\ (let raw_quads =\n FStar.Seq.Base.append #quad32 completed_quads (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0)\n inout_b) in let input_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads) 0 total_bytes in let padded_bytes =\n Vale.AES.GCTR_s.pad_to_128_bits input_bytes in let input_quads =\n Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in l_and (FStar.Seq.Base.length\n #Vale.Def.Types_s.quad32 input_quads > 0) (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8\n va_sM) == Vale.AES.GHash.ghash_incremental h_LE old_hash input_quads))) /\\ va_state_eq va_sM\n (va_update_flags va_sM (va_update_mem_heaplet 5 va_sM (va_update_xmm 10 va_sM (va_update_xmm 8\n va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5 va_sM (va_update_xmm 4\n va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1 va_sM (va_update_xmm 0\n va_sM (va_update_reg64 rR11 va_sM (va_update_reg64 rRcx va_sM (va_update_ok va_sM\n (va_update_mem va_sM va_s0))))))))))))))))))\n[@\"opaque_to_smt\"]\nlet va_lemma_Gcm_extra_bytes va_b0 va_s0 alg inout_b key round_keys keys_b hkeys_b total_bytes\n old_hash completed_quads h_LE =\n let (va_mods:va_mods_t) = [va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8;\n va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm\n 1; va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_ok; va_Mod_mem] in\n let va_qc = va_qcode_Gcm_extra_bytes va_mods alg inout_b key round_keys keys_b hkeys_b\n total_bytes old_hash completed_quads h_LE in\n let (va_sM, va_fM, va_g) = va_wp_sound_code_norm (va_code_Gcm_extra_bytes alg) va_qc va_s0 (fun\n va_s0 va_sM va_g -> let () = va_g in label va_range1\n \"***** POSTCONDITION NOT MET AT line 121 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\ (let (len:(va_int_range 1 1)) = 1 in label va_range1\n \"***** POSTCONDITION NOT MET AT line 174 column 55 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 177 column 95 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.AES.GCTR.gctr_partial alg len (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM) inout_b) key (va_get_xmm 11 va_s0)) /\\ label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 180 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let raw_quads = FStar.Seq.Base.append #quad32 completed_quads (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 181 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let input_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads) 0 total_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 182 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let padded_bytes = Vale.AES.GCTR_s.pad_to_128_bits input_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 183 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let input_quads = Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 186 column 59 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (l_and (FStar.Seq.Base.length #Vale.Def.Types_s.quad32 input_quads > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE old_hash input_quads)))))))) in\n assert_norm (va_qc.mods == va_mods);\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm\n 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1;\n va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_ok; va_Mod_mem]) va_sM va_s0;\n (va_sM, va_fM)\n\n[@ va_qattr]\nlet va_wp_Gcm_extra_bytes (alg:algorithm) (inout_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (total_bytes:nat) (old_hash:quad32)\n (completed_quads:(seq quad32)) (h_LE:quad32) (va_s0:va_state) (va_k:(va_state -> unit -> Type0))\n : Type0 =\n (va_get_ok va_s0 /\\ (let (len:(va_int_range 1 1)) = 1 in sse_enabled /\\\n Vale.X64.Decls.buffers_disjoint128 keys_b inout_b /\\ Vale.X64.Decls.buffers_disjoint128 hkeys_b\n inout_b /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) (va_get_reg64 rRax\n va_s0) inout_b len (va_get_mem_layout va_s0) Secret /\\ len == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 inout_b /\\ va_get_xmm 9 va_s0 == Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\ aes_reqs alg key round_keys\n keys_b (va_get_reg64 rR8 va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\\n pclmulqdq_enabled /\\ Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0\n va_s0) hkeys_b) (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret /\\ va_get_xmm 8 va_s0 == Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.AES.GHash.ghash_incremental0 h_LE old_hash completed_quads) /\\ FStar.Seq.Base.length\n #quad32 completed_quads == total_bytes `op_Division` 16 /\\ total_bytes < 16 `op_Multiply`\n FStar.Seq.Base.length #quad32 completed_quads + 16 /\\ va_get_reg64 rR10 va_s0 == total_bytes\n `op_Modulus` 16 /\\ total_bytes `op_Modulus` 16 =!= 0 /\\ (0 < total_bytes /\\ total_bytes < 16\n `op_Multiply` Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes) /\\ 16 `op_Multiply`\n (Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes - 1) < total_bytes) /\\ (forall\n (va_x_mem:vale_heap) (va_x_rcx:nat64) (va_x_r11:nat64) (va_x_xmm0:quad32) (va_x_xmm1:quad32)\n (va_x_xmm2:quad32) (va_x_xmm3:quad32) (va_x_xmm4:quad32) (va_x_xmm5:quad32) (va_x_xmm6:quad32)\n (va_x_xmm7:quad32) (va_x_xmm8:quad32) (va_x_xmm10:quad32) (va_x_heap5:vale_heap)\n (va_x_efl:Vale.X64.Flags.t) . let va_sM = va_upd_flags va_x_efl (va_upd_mem_heaplet 5\n va_x_heap5 (va_upd_xmm 10 va_x_xmm10 (va_upd_xmm 8 va_x_xmm8 (va_upd_xmm 7 va_x_xmm7\n (va_upd_xmm 6 va_x_xmm6 (va_upd_xmm 5 va_x_xmm5 (va_upd_xmm 4 va_x_xmm4 (va_upd_xmm 3 va_x_xmm3\n (va_upd_xmm 2 va_x_xmm2 (va_upd_xmm 1 va_x_xmm1 (va_upd_xmm 0 va_x_xmm0 (va_upd_reg64 rR11\n va_x_r11 (va_upd_reg64 rRcx va_x_rcx (va_upd_mem va_x_mem va_s0)))))))))))))) in va_get_ok\n va_sM /\\ (let (len:(va_int_range 1 1)) = 1 in Vale.X64.Decls.modifies_buffer128 inout_b\n (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5 va_sM) /\\ Vale.AES.GCTR.gctr_partial alg len\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b) key (va_get_xmm 11 va_s0) /\\ (let raw_quads =\n FStar.Seq.Base.append #quad32 completed_quads (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0)\n inout_b) in let input_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads) 0 total_bytes in let padded_bytes =\n Vale.AES.GCTR_s.pad_to_128_bits input_bytes in let input_quads =\n Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in l_and (FStar.Seq.Base.length\n #Vale.Def.Types_s.quad32 input_quads > 0) (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8\n va_sM) == Vale.AES.GHash.ghash_incremental h_LE old_hash input_quads))) ==> va_k va_sM (())))\n\n\nval va_wpProof_Gcm_extra_bytes : alg:algorithm -> inout_b:buffer128 -> key:(seq nat32) ->\n round_keys:(seq quad32) -> keys_b:buffer128 -> hkeys_b:buffer128 -> total_bytes:nat ->\n old_hash:quad32 -> completed_quads:(seq quad32) -> h_LE:quad32 -> va_s0:va_state ->\n va_k:(va_state -> unit -> Type0)\n -> Ghost (va_state & va_fuel & unit)\n (requires (va_t_require va_s0 /\\ va_wp_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b\n total_bytes old_hash completed_quads h_LE va_s0 va_k))\n (ensures (fun (va_sM, va_f0, va_g) -> va_t_ensure (va_code_Gcm_extra_bytes alg) ([va_Mod_flags;\n va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5;\n va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR11;\n va_Mod_reg64 rRcx; va_Mod_mem]) va_s0 va_k ((va_sM, va_f0, va_g))))\n[@\"opaque_to_smt\"]\nlet va_wpProof_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b total_bytes old_hash\n completed_quads h_LE va_s0 va_k =\n let (va_sM, va_f0) = va_lemma_Gcm_extra_bytes (va_code_Gcm_extra_bytes alg) va_s0 alg inout_b key\n round_keys keys_b hkeys_b total_bytes old_hash completed_quads h_LE in\n va_lemma_upd_update va_sM;\n assert (va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 5 va_sM (va_update_xmm 10\n va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5\n va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1\n va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR11 va_sM (va_update_reg64 rRcx va_sM\n (va_update_ok va_sM (va_update_mem va_sM va_s0)))))))))))))))));\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm\n 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1;\n va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_mem]) va_sM va_s0;\n let va_g = () in\n (va_sM, va_f0, va_g)\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_quick_Gcm_extra_bytes (alg:algorithm) (inout_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (total_bytes:nat) (old_hash:quad32)\n (completed_quads:(seq quad32)) (h_LE:quad32) : (va_quickCode unit (va_code_Gcm_extra_bytes alg)) =\n (va_QProc (va_code_Gcm_extra_bytes alg) ([va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10;\n va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm\n 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_mem])\n (va_wp_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b total_bytes old_hash\n completed_quads h_LE) (va_wpProof_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b\n total_bytes old_hash completed_quads h_LE))\n//--\n//-- Gcm_blocks128\n\nval va_code_Gcm_blocks128 : alg:algorithm -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_blocks128 alg =\n (va_Block (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRbx) (va_op_opr64_reg64 rRdi))\n (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR12) (va_op_opr64_reg64 rRdx)) (va_CCons\n (va_code_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRax)) (va_CCons\n (va_code_Ghash_buffer ()) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRdi)\n (va_op_opr64_reg64 rRbx)) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRdx)\n (va_op_opr64_reg64 rR12)) (va_CCons (va_code_Gctr_blocks128 alg) (va_CNil ())))))))))\n\nval va_codegen_success_Gcm_blocks128 : alg:algorithm -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_blocks128 alg =\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rRbx) (va_op_opr64_reg64 rRdi))\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR12) (va_op_opr64_reg64 rRdx))\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRax))\n (va_pbool_and (va_codegen_success_Ghash_buffer ()) (va_pbool_and (va_codegen_success_Mov64\n (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRbx)) (va_pbool_and (va_codegen_success_Mov64\n (va_op_dst_opr64_reg64 rRdx) (va_op_opr64_reg64 rR12)) (va_pbool_and\n (va_codegen_success_Gctr_blocks128 alg) (va_ttrue ()))))))))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_blocks128 (va_mods:va_mods_t) (alg:algorithm) (in_b:buffer128) (out_b:buffer128)\n (key:(seq nat32)) (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (h_LE:quad32)\n : (va_quickCode unit (va_code_Gcm_blocks128 alg)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 274 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRbx) (va_op_opr64_reg64 rRdi)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 275 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR12) (va_op_opr64_reg64 rRdx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 276 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRax)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 277 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Ghash_buffer hkeys_b in_b h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8\n va_old_s))) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 278 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRbx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 279 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRdx) (va_op_opr64_reg64 rR12)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 280 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gctr_blocks128 alg in_b out_b key round_keys keys_b) (va_QEmpty (()))))))))))\n\n\nval va_lemma_Gcm_blocks128 : va_b0:va_code -> va_s0:va_state -> alg:algorithm -> in_b:buffer128 ->\n out_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128 -> h_LE:quad32\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_blocks128 alg) va_s0 /\\ va_get_ok va_s0 /\\\n (sse_enabled /\\ Vale.X64.Decls.buffers_disjoint128 keys_b out_b /\\\n Vale.X64.Decls.buffers_disjoint128 hkeys_b out_b /\\ (Vale.X64.Decls.buffers_disjoint128 in_b\n out_b \\/ in_b == out_b) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0)\n (va_get_reg64 rRax va_s0) in_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1 va_s0) (va_get_reg64 rRdi va_s0) out_b\n (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\ va_get_reg64 rRax va_s0 + 16\n `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply`\n va_get_reg64 rRdx va_s0 < pow2_64 /\\ l_and (Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 in_b == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out_b)\n (Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b < pow2_32) /\\ va_get_reg64 rRdx\n va_s0 == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b /\\ va_get_xmm 9 va_s0 ==\n Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\\n va_get_reg64 rRdx va_s0 < pow2_32 /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rR8\n va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret)))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (Vale.X64.Decls.modifies_buffer128 out_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM) /\\ Vale.AES.GCTR.gctr_partial alg (va_get_reg64 rRdx va_s0) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b) key\n (va_get_xmm 11 va_s0) /\\ va_get_xmm 11 va_sM == Vale.AES.GCTR.inc32lite (va_get_xmm 11 va_s0)\n (va_get_reg64 rRdx va_s0) /\\ (va_get_reg64 rRdx va_s0 == 0 ==> l_and (va_get_xmm 8 va_sM ==\n va_get_xmm 8 va_s0) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b ==\n Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) out_b)) /\\ (va_get_reg64 rRdx va_s0 > 0 ==>\n l_and (va_get_reg64 rRdx va_s0 <= FStar.Seq.Base.length #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) ==> FStar.Seq.Base.length\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.slice #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) 0 (va_get_reg64 rRdx va_s0)) > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s0)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b)))) /\\ va_state_eq va_sM (va_update_flags va_sM\n (va_update_mem_heaplet 1 va_sM (va_update_xmm 10 va_sM (va_update_xmm 11 va_sM (va_update_xmm 8\n va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5 va_sM (va_update_xmm 4\n va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1 va_sM (va_update_xmm 0\n va_sM (va_update_reg64 rR12 va_sM (va_update_reg64 rRdx va_sM (va_update_reg64 rR10 va_sM\n (va_update_reg64 rR11 va_sM (va_update_reg64 rRdi va_sM (va_update_reg64 rRbx va_sM\n (va_update_ok va_sM (va_update_mem va_sM va_s0)))))))))))))))))))))))\n[@\"opaque_to_smt\"]\nlet va_lemma_Gcm_blocks128 va_b0 va_s0 alg in_b out_b key round_keys keys_b hkeys_b h_LE =\n let (va_mods:va_mods_t) = [va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11;\n va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm\n 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10;\n va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_ok; va_Mod_mem] in\n let va_qc = va_qcode_Gcm_blocks128 va_mods alg in_b out_b key round_keys keys_b hkeys_b h_LE in\n let (va_sM, va_fM, va_g) = va_wp_sound_code_norm (va_code_Gcm_blocks128 alg) va_qc va_s0 (fun\n va_s0 va_sM va_g -> let () = va_g in label va_range1\n \"***** POSTCONDITION NOT MET AT line 210 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\ (label va_range1\n \"***** POSTCONDITION NOT MET AT line 255 column 53 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 261 column 95 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.AES.GCTR.gctr_partial alg (va_get_reg64 rRdx va_s0) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b) key\n (va_get_xmm 11 va_s0)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 262 column 45 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_xmm 11 va_sM == Vale.AES.GCTR.inc32lite (va_get_xmm 11 va_s0) (va_get_reg64 rRdx\n va_s0)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 265 column 93 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRdx va_s0 == 0 ==> l_and (va_get_xmm 8 va_sM == va_get_xmm 8 va_s0)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b == Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) out_b)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 267 column 131 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRdx va_s0 > 0 ==> l_and (va_get_reg64 rRdx va_s0 <= FStar.Seq.Base.length\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) ==>\n FStar.Seq.Base.length #Vale.X64.Decls.quad32 (FStar.Seq.Base.slice #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) 0 (va_get_reg64 rRdx va_s0)) > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s0)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b))))) in\n assert_norm (va_qc.mods == va_mods);\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11; va_Mod_xmm\n 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2;\n va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10;\n va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_ok; va_Mod_mem]) va_sM va_s0;\n (va_sM, va_fM)\n\n[@ va_qattr]\nlet va_wp_Gcm_blocks128 (alg:algorithm) (in_b:buffer128) (out_b:buffer128) (key:(seq nat32))\n (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (h_LE:quad32) (va_s0:va_state)\n (va_k:(va_state -> unit -> Type0)) : Type0 =\n (va_get_ok va_s0 /\\ (sse_enabled /\\ Vale.X64.Decls.buffers_disjoint128 keys_b out_b /\\\n Vale.X64.Decls.buffers_disjoint128 hkeys_b out_b /\\ (Vale.X64.Decls.buffers_disjoint128 in_b\n out_b \\/ in_b == out_b) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0)\n (va_get_reg64 rRax va_s0) in_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1 va_s0) (va_get_reg64 rRdi va_s0) out_b\n (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\ va_get_reg64 rRax va_s0 + 16\n `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply`\n va_get_reg64 rRdx va_s0 < pow2_64 /\\ l_and (Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 in_b == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out_b)\n (Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b < pow2_32) /\\ va_get_reg64 rRdx\n va_s0 == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b /\\ va_get_xmm 9 va_s0 ==\n Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\\n va_get_reg64 rRdx va_s0 < pow2_32 /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rR8\n va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret) /\\ (forall (va_x_mem:vale_heap) (va_x_rbx:nat64) (va_x_rdi:nat64) (va_x_r11:nat64)\n (va_x_r10:nat64) (va_x_rdx:nat64) (va_x_r12:nat64) (va_x_xmm0:quad32) (va_x_xmm1:quad32)\n (va_x_xmm2:quad32) (va_x_xmm3:quad32) (va_x_xmm4:quad32) (va_x_xmm5:quad32) (va_x_xmm6:quad32)\n (va_x_xmm7:quad32) (va_x_xmm8:quad32) (va_x_xmm11:quad32) (va_x_xmm10:quad32)\n (va_x_heap1:vale_heap) (va_x_efl:Vale.X64.Flags.t) . let va_sM = va_upd_flags va_x_efl\n (va_upd_mem_heaplet 1 va_x_heap1 (va_upd_xmm 10 va_x_xmm10 (va_upd_xmm 11 va_x_xmm11\n (va_upd_xmm 8 va_x_xmm8 (va_upd_xmm 7 va_x_xmm7 (va_upd_xmm 6 va_x_xmm6 (va_upd_xmm 5 va_x_xmm5\n (va_upd_xmm 4 va_x_xmm4 (va_upd_xmm 3 va_x_xmm3 (va_upd_xmm 2 va_x_xmm2 (va_upd_xmm 1 va_x_xmm1\n (va_upd_xmm 0 va_x_xmm0 (va_upd_reg64 rR12 va_x_r12 (va_upd_reg64 rRdx va_x_rdx (va_upd_reg64\n rR10 va_x_r10 (va_upd_reg64 rR11 va_x_r11 (va_upd_reg64 rRdi va_x_rdi (va_upd_reg64 rRbx\n va_x_rbx (va_upd_mem va_x_mem va_s0))))))))))))))))))) in va_get_ok va_sM /\\\n (Vale.X64.Decls.modifies_buffer128 out_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM) /\\ Vale.AES.GCTR.gctr_partial alg (va_get_reg64 rRdx va_s0) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b) key\n (va_get_xmm 11 va_s0) /\\ va_get_xmm 11 va_sM == Vale.AES.GCTR.inc32lite (va_get_xmm 11 va_s0)\n (va_get_reg64 rRdx va_s0) /\\ (va_get_reg64 rRdx va_s0 == 0 ==> l_and (va_get_xmm 8 va_sM ==\n va_get_xmm 8 va_s0) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b ==\n Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) out_b)) /\\ (va_get_reg64 rRdx va_s0 > 0 ==>\n l_and (va_get_reg64 rRdx va_s0 <= FStar.Seq.Base.length #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) ==> FStar.Seq.Base.length\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.slice #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) 0 (va_get_reg64 rRdx va_s0)) > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s0)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b)))) ==> va_k va_sM (())))\n\n\nval va_wpProof_Gcm_blocks128 : alg:algorithm -> in_b:buffer128 -> out_b:buffer128 -> key:(seq\n nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 -> hkeys_b:buffer128 -> h_LE:quad32 ->\n va_s0:va_state -> va_k:(va_state -> unit -> Type0)\n -> Ghost (va_state & va_fuel & unit)\n (requires (va_t_require va_s0 /\\ va_wp_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b\n h_LE va_s0 va_k))\n (ensures (fun (va_sM, va_f0, va_g) -> va_t_ensure (va_code_Gcm_blocks128 alg) ([va_Mod_flags;\n va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6;\n va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0;\n va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10; va_Mod_reg64 rR11; va_Mod_reg64 rRdi;\n va_Mod_reg64 rRbx; va_Mod_mem]) va_s0 va_k ((va_sM, va_f0, va_g))))\n[@\"opaque_to_smt\"]\nlet va_wpProof_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b h_LE va_s0 va_k =\n let (va_sM, va_f0) = va_lemma_Gcm_blocks128 (va_code_Gcm_blocks128 alg) va_s0 alg in_b out_b key\n round_keys keys_b hkeys_b h_LE in\n va_lemma_upd_update va_sM;\n assert (va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 1 va_sM (va_update_xmm 10\n va_sM (va_update_xmm 11 va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6\n va_sM (va_update_xmm 5 va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2\n va_sM (va_update_xmm 1 va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR12 va_sM\n (va_update_reg64 rRdx va_sM (va_update_reg64 rR10 va_sM (va_update_reg64 rR11 va_sM\n (va_update_reg64 rRdi va_sM (va_update_reg64 rRbx va_sM (va_update_ok va_sM (va_update_mem\n va_sM va_s0))))))))))))))))))))));\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11; va_Mod_xmm\n 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2;\n va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10;\n va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_mem]) va_sM va_s0;\n let va_g = () in\n (va_sM, va_f0, va_g)\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_quick_Gcm_blocks128 (alg:algorithm) (in_b:buffer128) (out_b:buffer128) (key:(seq nat32))\n (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (h_LE:quad32) : (va_quickCode\n unit (va_code_Gcm_blocks128 alg)) =\n (va_QProc (va_code_Gcm_blocks128 alg) ([va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10;\n va_Mod_xmm 11; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm\n 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64\n rR10; va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_mem])\n (va_wp_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b h_LE)\n (va_wpProof_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b h_LE))\n//--\n//-- Gcm_blocks\n#push-options \"--z3rlimit 1000\"\n\nval va_code_Gcm_blocks : alg:algorithm -> offset:int -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_blocks alg offset =\n (va_Block (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR13) (va_op_opr64_reg64 rRcx))\n (va_CCons (va_code_AddLea64 (va_op_dst_opr64_reg64 rR9) (va_op_opr64_reg64 rR9) (va_const_opr64\n 32)) (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rRbx) (va_op_reg_opr64_reg64 rRsp)\n (offset + 0)) (va_CCons (va_code_Gcm_blocks_auth ()) (va_CCons (va_code_Load64_stack\n (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 8)) (va_CCons\n (va_code_Load64_stack (va_op_dst_opr64_reg64 rRsi) (va_op_reg_opr64_reg64 rRsp) (offset + 16))\n (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp)\n (offset + 24)) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRcx) (va_op_opr64_reg64 rR13))\n (va_CCons (va_code_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (va_CCons\n (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 2) (va_op_xmm_xmm 1) (va_op_reg_opr64_reg64\n rR8) 0 Public) (va_CCons (va_code_Store128_buffer (va_op_heaplet_mem_heaplet 3)\n (va_op_reg_opr64_reg64 rRbp) (va_op_xmm_xmm 1) 0 Secret) (va_CCons (va_code_Load_one_lsb\n (va_op_xmm_xmm 10)) (va_CCons (va_code_VPaddd (va_op_xmm_xmm 1) (va_op_xmm_xmm 1)\n (va_op_xmm_xmm 10)) (va_CCons (va_code_AES_GCM_decrypt_6mult alg) (va_CCons\n (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 11) (va_op_reg_opr64_reg64\n rRbp) 32 Secret) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR8) (va_op_opr64_reg64 rRcx))\n (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp)\n (offset + 32)) (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rRdi)\n (va_op_reg_opr64_reg64 rRsp) (offset + 40)) (va_CCons (va_code_Load64_stack\n (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp) (offset + 48)) (va_CCons\n (va_code_Mov64 (va_op_dst_opr64_reg64 rR14) (va_op_opr64_reg64 rRdx)) (va_CCons\n (va_code_InitPshufbMask (va_op_xmm_xmm 9) (va_op_reg_opr64_reg64 rR12)) (va_CCons\n (va_code_Pshufb (va_op_xmm_xmm 11) (va_op_xmm_xmm 9)) (va_CCons (va_code_Gcm_blocks128 alg)\n (va_CCons (va_code_Stack_lemma ()) (va_CCons (va_code_Add64 (va_op_dst_opr64_reg64 rR14)\n (va_opr_code_Stack (va_op_reg64_reg64 rRsp) (offset + 24) Public)) (va_CCons (va_code_IMul64\n (va_op_dst_opr64_reg64 rR14) (va_const_opr64 16)) (va_CCons (va_code_Load64_stack\n (va_op_dst_opr64_reg64 rR13) (va_op_reg_opr64_reg64 rRsp) (offset + 64)) (va_CCons (va_IfElse\n (va_cmp_gt (va_op_cmp_reg64 rR13) (va_op_cmp_reg64 rR14)) (va_Block (va_CCons\n (va_code_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp) (offset + 56))\n (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR10) (va_op_opr64_reg64 rR13)) (va_CCons\n (va_code_And64 (va_op_dst_opr64_reg64 rR10) (va_const_opr64 15)) (va_CCons\n (va_code_Gcm_extra_bytes alg) (va_CCons (va_Block (va_CNil ())) (va_CNil ()))))))) (va_Block\n (va_CNil ()))) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR11) (va_op_opr64_reg64 rR15))\n (va_CCons (va_code_Gcm_make_length_quad ()) (va_CCons (va_code_Ghash_register ()) (va_CCons\n (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64\n rRbp) 0 Secret) (va_CCons (va_code_Gctr_register alg) (va_CCons (va_Block (va_CNil ()))\n (va_CNil ()))))))))))))))))))))))))))))))))))))\n\nval va_codegen_success_Gcm_blocks : alg:algorithm -> offset:int -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_blocks alg offset =\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR13) (va_op_opr64_reg64 rRcx))\n (va_pbool_and (va_codegen_success_AddLea64 (va_op_dst_opr64_reg64 rR9) (va_op_opr64_reg64 rR9)\n (va_const_opr64 32)) (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64\n rRbx) (va_op_reg_opr64_reg64 rRsp) (offset + 0)) (va_pbool_and\n (va_codegen_success_Gcm_blocks_auth ()) (va_pbool_and (va_codegen_success_Load64_stack\n (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 8)) (va_pbool_and\n (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRsi) (va_op_reg_opr64_reg64 rRsp)\n (offset + 16)) (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRdx)\n (va_op_reg_opr64_reg64 rRsp) (offset + 24)) (va_pbool_and (va_codegen_success_Mov64\n (va_op_dst_opr64_reg64 rRcx) (va_op_opr64_reg64 rR13)) (va_pbool_and (va_codegen_success_Mov128\n (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (va_pbool_and (va_codegen_success_Load128_buffer\n (va_op_heaplet_mem_heaplet 2) (va_op_xmm_xmm 1) (va_op_reg_opr64_reg64 rR8) 0 Public)\n (va_pbool_and (va_codegen_success_Store128_buffer (va_op_heaplet_mem_heaplet 3)\n (va_op_reg_opr64_reg64 rRbp) (va_op_xmm_xmm 1) 0 Secret) (va_pbool_and\n (va_codegen_success_Load_one_lsb (va_op_xmm_xmm 10)) (va_pbool_and (va_codegen_success_VPaddd\n (va_op_xmm_xmm 1) (va_op_xmm_xmm 1) (va_op_xmm_xmm 10)) (va_pbool_and\n (va_codegen_success_AES_GCM_decrypt_6mult alg) (va_pbool_and (va_codegen_success_Load128_buffer\n (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 11) (va_op_reg_opr64_reg64 rRbp) 32 Secret)\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR8) (va_op_opr64_reg64 rRcx))\n (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRax)\n (va_op_reg_opr64_reg64 rRsp) (offset + 32)) (va_pbool_and (va_codegen_success_Load64_stack\n (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 40)) (va_pbool_and\n (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp)\n (offset + 48)) (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR14)\n (va_op_opr64_reg64 rRdx)) (va_pbool_and (va_codegen_success_InitPshufbMask (va_op_xmm_xmm 9)\n (va_op_reg_opr64_reg64 rR12)) (va_pbool_and (va_codegen_success_Pshufb (va_op_xmm_xmm 11)\n (va_op_xmm_xmm 9)) (va_pbool_and (va_codegen_success_Gcm_blocks128 alg) (va_pbool_and\n (va_codegen_success_Stack_lemma ()) (va_pbool_and (va_codegen_success_Add64\n (va_op_dst_opr64_reg64 rR14) (va_opr_code_Stack (va_op_reg64_reg64 rRsp) (offset + 24) Public))\n (va_pbool_and (va_codegen_success_IMul64 (va_op_dst_opr64_reg64 rR14) (va_const_opr64 16))\n (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rR13)\n (va_op_reg_opr64_reg64 rRsp) (offset + 64)) (va_pbool_and (va_pbool_and\n (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp)\n (offset + 56)) (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR10)\n (va_op_opr64_reg64 rR13)) (va_pbool_and (va_codegen_success_And64 (va_op_dst_opr64_reg64 rR10)\n (va_const_opr64 15)) (va_codegen_success_Gcm_extra_bytes alg)))) (va_pbool_and\n (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR11) (va_op_opr64_reg64 rR15)) (va_pbool_and\n (va_codegen_success_Gcm_make_length_quad ()) (va_pbool_and (va_codegen_success_Ghash_register\n ()) (va_pbool_and (va_codegen_success_Load128_buffer (va_op_heaplet_mem_heaplet 3)\n (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64 rRbp) 0 Secret) (va_pbool_and\n (va_codegen_success_Gctr_register alg) (va_ttrue ()))))))))))))))))))))))))))))))))))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_blocks (va_mods:va_mods_t) (alg:algorithm) (offset:int) (auth_b:buffer128)\n (abytes_b:buffer128) (in128x6_b:buffer128) (out128x6_b:buffer128) (in128_b:buffer128)\n (out128_b:buffer128) (inout_b:buffer128) (iv_b:buffer128) (scratch_b:buffer128) (key:(seq nat32))\n (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) : (va_quickCode unit\n (va_code_Gcm_blocks alg offset)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in let\n (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 0) (va_get_stack va_s) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 8) (va_get_stack va_s) in let\n (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 16) (va_get_stack va_s) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 24) (va_get_stack va_s) in let\n (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 32) (va_get_stack va_s) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 40) (va_get_stack va_s) in let\n (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset\n + 48) (va_get_stack va_s) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 56) (va_get_stack va_s) in let\n (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s\n + offset + 64) (va_get_stack va_s) in let (h_LE:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.X64.Decls.buffer128_read hkeys_b 2\n (va_get_mem_heaplet 0 va_old_s)) in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 463 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR13) (va_op_opr64_reg64 rRcx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 464 column 13 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_AddLea64 (va_op_dst_opr64_reg64 rR9) (va_op_opr64_reg64 rR9) (va_const_opr64 32))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 465 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRbx) (va_op_reg_opr64_reg64 rRsp) (offset + 0))\n (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 466 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_blocks_auth auth_b abytes_b hkeys_b h_LE) (fun (va_s:va_state)\n (auth_quad_seq:(seq quad32)) -> let (y_0:quad32) = Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 0 0 0 0 in let (y_auth_bytes:quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s) in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 473 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 8))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 474 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRsi) (va_op_reg_opr64_reg64 rRsp) (offset + 16))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 475 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp) (offset + 24))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 476 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRcx) (va_op_opr64_reg64 rR13)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 477 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (fun (va_s:va_state) _ -> let\n (iv_BE:Vale.X64.Decls.quad32) = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2\n va_old_s) in let (ctr_BE_1:quad32) = iv_BE in let (ctr_BE_2:quad32) = Vale.AES.GCTR_s.inc32\n iv_BE 1 in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 483 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 2) (va_op_xmm_xmm 1) (va_op_reg_opr64_reg64\n rR8) 0 Public iv_b 0) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 485 column 20 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Store128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_reg_opr64_reg64 rRbp)\n (va_op_xmm_xmm 1) 0 Secret scratch_b 0) (fun (va_s:va_state) _ -> let (j0:quad32) = va_get_xmm\n 1 va_s in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 487 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load_one_lsb (va_op_xmm_xmm 10)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 489 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_VPaddd (va_op_xmm_xmm 1) (va_op_xmm_xmm 1) (va_op_xmm_xmm 10)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 491 column 26 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_AES_GCM_decrypt_6mult alg h_LE iv_b in128x6_b out128x6_b scratch_b key round_keys\n keys_b hkeys_b) (fun (va_s:va_state) _ -> let (y_cipher128x6:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s) in let (auth_in:(seq quad32)) =\n auth_quad_seq in let (va_arg138:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b in let\n (va_arg137:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = auth_in in let\n (va_arg136:Vale.Def.Types_s.quad32) = y_auth_bytes in let (va_arg135:Vale.Def.Types_s.quad32) =\n y_0 in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 494 column 36 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GHash.lemma_ghash_incremental0_append h_LE va_arg135 va_arg136\n y_cipher128x6 va_arg137 va_arg138) (let auth_in = FStar.Seq.Base.append #quad32 auth_in\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b) in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 498 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 11)\n (va_op_reg_opr64_reg64 rRbp) 32 Secret scratch_b 2) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 499 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR8) (va_op_opr64_reg64 rRcx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 500 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp) (offset + 32))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 501 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 40))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 502 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp) (offset + 48))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 503 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR14) (va_op_opr64_reg64 rRdx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 504 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_InitPshufbMask (va_op_xmm_xmm 9) (va_op_reg_opr64_reg64 rR12)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 505 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Pshufb (va_op_xmm_xmm 11) (va_op_xmm_xmm 9)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 506 column 18 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_blocks128 alg in128_b out128_b key round_keys keys_b hkeys_b h_LE) (fun\n (va_s:va_state) _ -> let (y_cipher128:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s) in let (va_arg134:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b in let\n (va_arg133:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = auth_in in let\n (va_arg132:Vale.Def.Types_s.quad32) = y_0 in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 508 column 36 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GHash.lemma_ghash_incremental0_append h_LE va_arg132 y_cipher128x6\n y_cipher128 va_arg133 va_arg134) (let auth_in = FStar.Seq.Base.append #quad32 auth_in\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b) in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 512 column 22 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Stack_lemma (va_op_reg64_reg64 rRsp) (offset + 24) Public) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 512 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Add64 (va_op_dst_opr64_reg64 rR14) (va_opr_code_Stack (va_op_reg64_reg64 rRsp)\n (offset + 24) Public)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 513 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_IMul64 (va_op_dst_opr64_reg64 rR14) (va_const_opr64 16)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 514 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rR13) (va_op_reg_opr64_reg64 rRsp) (offset + 64))\n (fun (va_s:va_state) _ -> let (y_inout:Vale.Def.Types_s.quad32) = y_cipher128 in let\n (plain_byte_seq:(seq quad32)) = empty_seq_quad32 in let (cipher_byte_seq:(seq quad32)) =\n empty_seq_quad32 in let (va_arg131:Vale.Def.Types_s.quad32) = va_get_xmm 11 va_s in let\n (va_arg130:(FStar.Seq.Base.seq Vale.Def.Types_s.nat32)) = key in let\n (va_arg129:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = cipher_byte_seq in let\n (va_arg128:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = plain_byte_seq in let\n (va_arg127:Vale.AES.AES_common_s.algorithm) = alg in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 519 column 29 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.gctr_partial_opaque_init va_arg127 va_arg128 va_arg129 va_arg130\n va_arg131) (let (total_bytes:(va_int_at_least 0)) = FStar.Seq.Base.length #quad32 auth_quad_seq\n `op_Multiply` 16 + plain_num_bytes in va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 523 column 8 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_qIf va_mods (Cmp_gt (va_op_cmp_reg64 rR13) (va_op_cmp_reg64 rR14)) (qblock va_mods (fun\n (va_s:va_state) -> va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 525 column 21 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp) (offset + 56))\n (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 526 column 14 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR10) (va_op_opr64_reg64 rR13)) (fun (va_s:va_state) _\n -> va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 527 column 26 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.Poly1305.Math.lemma_poly_bits64 ()) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 528 column 14 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_And64 (va_op_dst_opr64_reg64 rR10) (va_const_opr64 15)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 532 column 24 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b total_bytes y_0 auth_in\n h_LE) (fun (va_s:va_state) _ -> let y_inout = Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm\n 8 va_s) in let (raw_auth_quads:(FStar.Seq.Base.seq quad32)) = FStar.Seq.Base.append #quad32\n auth_in (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_old_s) inout_b) in va_qAssertSquash\n va_range1\n \"***** EXPRESSION PRECONDITIONS NOT MET WITHIN line 536 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n ((fun a_1906 (s_1907:(FStar.Seq.Base.seq a_1906)) (i_1908:Prims.nat) (j_1909:Prims.nat) -> let\n (j_1869:Prims.nat) = j_1909 in Prims.b2t (Prims.op_AmpAmp (Prims.op_LessThanOrEqual i_1908\n j_1869) (Prims.op_LessThanOrEqual j_1869 (FStar.Seq.Base.length #a_1906 s_1907))))\n Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_auth_quads) 0 total_bytes)\n (fun _ -> let (auth_input_bytes:(FStar.Seq.Base.seq Vale.Def.Types_s.nat8)) =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n raw_auth_quads) 0 total_bytes in let (padded_auth_bytes:(FStar.Seq.Base.seq\n Vale.Def.Types_s.nat8)) = Vale.AES.GCTR_s.pad_to_128_bits auth_input_bytes in let auth_in =\n Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_auth_bytes in let plain_byte_seq =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_old_s) inout_b in let cipher_byte_seq =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s) inout_b in va_QEmpty ((auth_in,\n cipher_byte_seq, plain_byte_seq, y_inout)))))))))) (qblock va_mods (fun (va_s:va_state) ->\n va_QEmpty ((auth_in, cipher_byte_seq, plain_byte_seq, y_inout))))) (fun (va_s:va_state) va_g ->\n let ((auth_in:(seq quad32)), (cipher_byte_seq:(seq quad32)), (plain_byte_seq:(seq quad32)),\n (y_inout:Vale.Def.Types_s.quad32)) = va_g in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 547 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR11) (va_op_opr64_reg64 rR15)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 548 column 25 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_make_length_quad ()) (fun (va_s:va_state) _ -> let\n (length_quad32:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 0\n va_s) in va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 551 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Ghash_register hkeys_b h_LE y_inout) (fun (va_s:va_state) _ -> let\n (y_final:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s)\n in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 554 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64\n rRbp) 0 Secret scratch_b 0) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 557 column 18 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gctr_register alg key round_keys keys_b) (fun (va_s:va_state) _ -> let\n (va_arg126:Vale.Def.Types_s.quad32) = va_get_xmm 8 va_s in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 560 column 40 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.Arch.Types.le_seq_quad32_to_bytes_of_singleton va_arg126)\n (va_qAssertSquash va_range1\n \"***** EXPRESSION PRECONDITIONS NOT MET WITHIN line 561 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n ((fun (icb_BE_677:Vale.Def.Types_s.quad32) (plain_LE_678:Vale.Def.Types_s.quad32)\n (alg_679:Vale.AES.AES_common_s.algorithm) (key_680:(FStar.Seq.Base.seq Vale.Def.Types_s.nat32))\n (i_681:Prims.int) -> Vale.AES.AES_s.is_aes_key_LE alg_679 key_680) j0 y_final alg key 0) (fun _\n -> va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 561 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_xmm 8 va_s == Vale.AES.GCTR_s.gctr_encrypt_block j0 y_final alg key 0) (let\n (plain128:(FStar.Seq.Base.seq Vale.X64.Decls.quad32)) = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b) in let\n (cipher128:(FStar.Seq.Base.seq Vale.X64.Decls.quad32)) = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) in128_b) in va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 566 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (FStar.Seq.Base.length #quad32 plain_byte_seq == 0 ==> FStar.Seq.Base.equal\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 plain128 plain_byte_seq)\n plain128) (va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 567 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (FStar.Seq.Base.length #quad32 cipher_byte_seq == 0 ==> FStar.Seq.Base.equal\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 cipher128 cipher_byte_seq)\n cipher128) (let (va_arg125:Vale.Def.Types_s.quad32) = Vale.AES.GCTR.inc32lite ctr_BE_2 len128x6\n in let (va_arg124:Vale.Def.Types_s.quad32) = ctr_BE_2 in let (va_arg123:(FStar.Seq.Base.seq\n Vale.Def.Types_s.nat32)) = key in let (va_arg122:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32))\n = Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) out128_b in let\n (va_arg121:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_old_s) in128_b in let (va_arg120:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s) out128x6_b in let\n (va_arg119:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_old_s) in128x6_b in let (va_arg118:Prims.nat) = len128 in let\n (va_arg117:Prims.nat) = len128x6 in let (va_arg116:Vale.AES.AES_common_s.algorithm) = alg in\n va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 569 column 30 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.lemma_gctr_partial_append va_arg116 va_arg117 va_arg118\n va_arg119 va_arg120 va_arg121 va_arg122 va_arg123 va_arg124 va_arg125) (let\n (va_arg115:Vale.Def.Types_s.quad32) = Vale.AES.GCTR.inc32lite (Vale.AES.GCTR.inc32lite ctr_BE_2\n len128x6) len128 in let (va_arg114:Vale.Def.Types_s.quad32) = ctr_BE_2 in let\n (va_arg113:(FStar.Seq.Base.seq Vale.Def.Types_s.nat32)) = key in let\n (va_arg112:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = cipher_byte_seq in let\n (va_arg111:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = plain_byte_seq in let\n (va_arg110:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s) out128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) out128_b) in let\n (va_arg109:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b) in let (va_arg108:Prims.nat) =\n FStar.Seq.Base.length #quad32 plain_byte_seq in let (va_arg107:Prims.nat) = len128x6 + len128\n in let (va_arg106:Vale.AES.AES_common_s.algorithm) = alg in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 575 column 30 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.lemma_gctr_partial_append va_arg106 va_arg107 va_arg108\n va_arg109 va_arg110 va_arg111 va_arg112 va_arg113 va_arg114 va_arg115) (let\n (va_arg105:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = auth_in in let\n (va_arg104:Vale.Def.Types_s.quad32) = y_0 in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 583 column 23 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GHash.lemma_hash_append2 h_LE va_arg104 y_inout y_final va_arg105\n length_quad32) (let auth_in = FStar.Seq.Base.append #quad32 auth_in (FStar.Seq.Base.create\n #Vale.Def.Types_s.quad32 1 length_quad32) in let (va_arg103:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = auth_in in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 585 column 31 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GHash.ghash_incremental_to_ghash h_LE va_arg103) (va_QEmpty\n (()))))))))))))))))))))))))))))))))))))))))))))))))\n\n\nval va_lemma_Gcm_blocks : va_b0:va_code -> va_s0:va_state -> alg:algorithm -> offset:int ->\n auth_b:buffer128 -> abytes_b:buffer128 -> in128x6_b:buffer128 -> out128x6_b:buffer128 ->\n in128_b:buffer128 -> out128_b:buffer128 -> inout_b:buffer128 -> iv_b:buffer128 ->\n scratch_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_blocks alg offset) va_s0 /\\ va_get_ok va_s0 /\\\n (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in let (h_LE:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.X64.Decls.buffer128_read hkeys_b 2\n (va_get_mem_heaplet 0 va_s0)) in sse_enabled /\\ movbe_enabled /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0)\n Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp\n va_s0 + offset + 8) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 24) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 40) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 56) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet\n 1 va_s0) (va_get_reg64 rRdi va_s0) auth_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0)\n Secret /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 7 va_s0) abytes_ptr abytes_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 2\n va_s0) (va_get_reg64 rR8 va_s0) iv_b 1 (va_get_mem_layout va_s0) Public /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 6 va_s0) in128x6_ptr in128x6_b len128x6\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 6\n va_s0) out128x6_ptr out128x6_b len128x6 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0) in128_ptr in128_b len128\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1\n va_s0) out128_ptr out128_b len128 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) inout_ptr inout_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 3\n va_s0) (va_get_reg64 rRbp va_s0) scratch_b 9 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0) hkeys_b 8\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.buffer_disjoints128 iv_b ([keys_b;\n scratch_b; in128x6_b; out128x6_b; hkeys_b; in128_b; out128_b; inout_b]) /\\\n Vale.X64.Decls.buffer_disjoints128 scratch_b ([keys_b; in128x6_b; out128x6_b; in128_b;\n out128_b; inout_b; hkeys_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128x6_b ([keys_b;\n hkeys_b; in128_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128_b ([keys_b; hkeys_b;\n out128x6_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 inout_b ([keys_b; hkeys_b;\n out128x6_b; out128_b]) /\\ (Vale.X64.Decls.buffers_disjoint128 in128x6_b out128x6_b \\/ in128x6_b\n == out128x6_b) /\\ (Vale.X64.Decls.buffers_disjoint128 in128_b out128_b \\/ in128_b == out128_b)\n /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ in128x6_ptr\n + 16 `op_Multiply` len128x6 < pow2_64 /\\ out128x6_ptr + 16 `op_Multiply` len128x6 < pow2_64 /\\\n in128_ptr + 16 `op_Multiply` len128 < pow2_64 /\\ out128_ptr + 16 `op_Multiply` len128 < pow2_64\n /\\ inout_ptr + 16 < pow2_64 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 auth_b ==\n va_get_reg64 rRdx va_s0 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 abytes_b == 1\n /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128x6_b ==\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out128x6_b /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 out128_b /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128\n in128x6_b == len128x6 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b ==\n len128 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 inout_b == 1 /\\\n plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32 /\\ va_get_reg64 rR9 va_s0 + 32 <\n pow2_64 /\\ Vale.X64.Memory.buffer_addr #Vale.X64.Memory.vuint128 keys_b (va_get_mem_heaplet 0\n va_s0) + 128 < pow2_64 /\\ len128x6 `op_Modulus` 6 == 0 /\\ (len128x6 > 0 ==> len128x6 >= 6) /\\\n 12 + len128x6 + 6 < pow2_32 /\\ (va_mul_nat len128x6 (128 `op_Division` 8) + va_mul_nat len128\n (128 `op_Division` 8) <= plain_num_bytes /\\ plain_num_bytes < va_mul_nat len128x6 (128\n `op_Division` 8) + va_mul_nat len128 (128 `op_Division` 8) + 128 `op_Division` 8) /\\\n (va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) <= va_get_reg64 rRsi va_s0 /\\\n va_get_reg64 rRsi va_s0 < va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) + 128\n `op_Division` 8) /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rRcx va_s0)\n (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 (Vale.AES.AES_s.aes_encrypt_LE alg key\n (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0))))))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in let (h_LE:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.X64.Decls.buffer128_read hkeys_b 2\n (va_get_mem_heaplet 0 va_s0)) in Vale.X64.Decls.modifies_buffer128 out128_b (va_get_mem_heaplet\n 1 va_s0) (va_get_mem_heaplet 1 va_sM) /\\ Vale.X64.Decls.modifies_buffer128 iv_b\n (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2 va_sM) /\\ Vale.X64.Decls.modifies_buffer128\n scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3 va_sM) /\\\n Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM) /\\ Vale.X64.Decls.modifies_buffer128 out128x6_b (va_get_mem_heaplet 6 va_s0)\n (va_get_mem_heaplet 6 va_sM) /\\ plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32\n /\\ (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in let\n (ctr_BE_1:quad32) = iv_BE in let (ctr_BE_2:quad32) = Vale.AES.GCTR_s.inc32 iv_BE 1 in let\n (plain_in:(seq quad32)) = (if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) then FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b) else FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in128_b)) in let (cipher_out:(seq quad32)) = (if (plain_num_bytes\n > (len128x6 + len128) `op_Multiply` 128 `op_Division` 8) then FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM)\n out128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM) inout_b) else\n FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM)\n out128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out128_b)) in let\n (cipher_bound:nat) = (if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128 `op_Division`\n 8) then (len128x6 + len128 + 1) else (len128x6 + len128)) in Vale.AES.GCTR.gctr_partial alg\n cipher_bound plain_in cipher_out key ctr_BE_2 /\\ (let (length_quad:quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.Def.Types_s.insert_nat64\n (Vale.Def.Types_s.insert_nat64 (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0) (8\n `op_Multiply` va_get_reg64 rRsi va_s0) 1) (8 `op_Multiply` plain_num_bytes) 0) in let\n (raw_auth_quads:(seq quad32)) = (if (va_get_reg64 rRsi va_s0 > va_get_reg64 rRdx va_s0\n `op_Multiply` 128 `op_Division` 8) then FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 7 va_s0) abytes_b) else Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0)\n auth_b) in let (auth_input_bytes:(seq nat8)) = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_auth_quads) 0 (va_get_reg64 rRsi va_s0) in let\n (padded_auth_bytes:(seq nat8)) = Vale.AES.GCTR_s.pad_to_128_bits auth_input_bytes in let\n (auth_quad_seq:(seq quad32)) = Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_auth_bytes in let\n (raw_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32 (FStar.Seq.Base.append #quad32\n auth_quad_seq (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b))\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b) in let (total_bytes:nat) =\n FStar.Seq.Base.length #quad32 auth_quad_seq `op_Multiply` 16 + plain_num_bytes in let\n (raw_quad_seq:(seq quad32)) = (if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) then (let (ab:(seq nat8)) = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes (FStar.Seq.Base.append #quad32 raw_quad_seq\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b))) 0 total_bytes in let (pb:(seq\n nat8)) = Vale.AES.GCTR_s.pad_to_128_bits ab in Vale.Def.Types_s.le_bytes_to_seq_quad32 pb) else\n raw_quad_seq) in let (auth_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32 raw_quad_seq\n (FStar.Seq.Base.create #quad32 1 length_quad) in va_get_xmm 8 va_sM ==\n Vale.AES.GCTR_s.gctr_encrypt_block ctr_BE_1 (Vale.AES.GHash_s.ghash_LE h_LE auth_quad_seq) alg\n key 0))) /\\ va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 6 va_sM\n (va_update_mem_heaplet 5 va_sM (va_update_mem_heaplet 3 va_sM (va_update_mem_heaplet 2 va_sM\n (va_update_mem_heaplet 1 va_sM (va_update_xmm 15 va_sM (va_update_xmm 14 va_sM (va_update_xmm\n 13 va_sM (va_update_xmm 12 va_sM (va_update_xmm 11 va_sM (va_update_xmm 10 va_sM (va_update_xmm\n 9 va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5\n va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1\n va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR15 va_sM (va_update_reg64 rR14 va_sM\n (va_update_reg64 rR13 va_sM (va_update_reg64 rR12 va_sM (va_update_reg64 rR11 va_sM\n (va_update_reg64 rR10 va_sM (va_update_reg64 rR9 va_sM (va_update_reg64 rR8 va_sM\n (va_update_reg64 rRbp va_sM (va_update_reg64 rRsi va_sM (va_update_reg64 rRdi va_sM\n (va_update_reg64 rRdx va_sM (va_update_reg64 rRcx va_sM (va_update_reg64 rRbx va_sM\n (va_update_reg64 rRax va_sM (va_update_ok va_sM (va_update_mem va_sM\n va_s0)))))))))))))))))))))))))))))))))))))))))\n[@\"opaque_to_smt\"]\nlet va_lemma_Gcm_blocks va_b0 va_s0 alg offset auth_b abytes_b in128x6_b out128x6_b in128_b\n out128_b inout_b iv_b scratch_b key round_keys keys_b hkeys_b =\n let (va_mods:va_mods_t) = [va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5;\n va_Mod_mem_heaplet 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14;\n va_Mod_xmm 13; va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8;\n va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm\n 1; va_Mod_xmm 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12;\n va_Mod_reg64 rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp;\n va_Mod_reg64 rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx;\n va_Mod_reg64 rRax; va_Mod_ok; va_Mod_mem] in\n let va_qc = va_qcode_Gcm_blocks va_mods alg offset auth_b abytes_b in128x6_b out128x6_b in128_b\n out128_b inout_b iv_b scratch_b key round_keys keys_b hkeys_b in\n let (va_sM, va_fM, va_g) = va_wp_sound_code_norm (va_code_Gcm_blocks alg offset) va_qc va_s0 (fun\n va_s0 va_sM va_g -> let () = va_g in label va_range1\n \"***** POSTCONDITION NOT MET AT line 283 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\ (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64\n (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0) in let\n (in128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 8) (va_get_stack va_s0) in let (out128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack va_s0) in\n let (len128x6:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 24) (va_get_stack va_s0) in let (in128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack va_s0) in\n let (out128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 40) (va_get_stack va_s0) in let (len128:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack va_s0) in\n let (inout_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 56) (va_get_stack va_s0) in let (plain_num_bytes:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack va_s0) in\n let (h_LE:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.X64.Decls.buffer128_read hkeys_b 2 (va_get_mem_heaplet 0 va_s0)) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 396 column 56 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out128_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 397 column 52 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 iv_b (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 398 column 57 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 399 column 55 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 400 column 58 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out128x6_b (va_get_mem_heaplet 6 va_s0) (va_get_mem_heaplet\n 6 va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 403 column 39 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (plain_num_bytes < pow2_32) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 404 column 38 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRsi va_s0 < pow2_32) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 406 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 408 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (ctr_BE_1:quad32) = iv_BE in label va_range1\n \"***** POSTCONDITION NOT MET AT line 409 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (ctr_BE_2:quad32) = Vale.AES.GCTR_s.inc32 iv_BE 1 in label va_range1\n \"***** POSTCONDITION NOT MET AT line 412 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (plain_in:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b)) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in128_b)) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 421 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (cipher_out:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM) out128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b)) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_sM) out128_b)) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 430 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (cipher_bound:nat) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> len128x6 + len128 + 1) (fun _ -> len128x6 + len128) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 434 column 77 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.AES.GCTR.gctr_partial alg cipher_bound plain_in cipher_out key ctr_BE_2) /\\ label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 438 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (length_quad:quad32) = Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.Def.Types_s.insert_nat64 (Vale.Def.Types_s.insert_nat64 (Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 0 0 0 0) (8 `op_Multiply` va_get_reg64 rRsi va_s0) 1) (8 `op_Multiply`\n plain_num_bytes) 0) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 440 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (raw_auth_quads:(seq quad32)) = va_if (va_get_reg64 rRsi va_s0 > va_get_reg64 rRdx va_s0\n `op_Multiply` 128 `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 7 va_s0) abytes_b)) (fun _ -> Vale.X64.Decls.s128 (va_get_mem_heaplet 1\n va_s0) auth_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 444 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (auth_input_bytes:(seq nat8)) = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_auth_quads) 0 (va_get_reg64 rRsi va_s0) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 445 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (padded_auth_bytes:(seq nat8)) = Vale.AES.GCTR_s.pad_to_128_bits auth_input_bytes in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 446 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (auth_quad_seq:(seq quad32)) = Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_auth_bytes\n in label va_range1\n \"***** POSTCONDITION NOT MET AT line 448 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (raw_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32 (FStar.Seq.Base.append #quad32\n auth_quad_seq (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b))\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 452 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (total_bytes:nat) = FStar.Seq.Base.length #quad32 auth_quad_seq `op_Multiply` 16 +\n plain_num_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 453 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (raw_quad_seq:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply`\n 128 `op_Division` 8) (fun _ -> let (ab:(seq nat8)) = FStar.Seq.Base.slice\n #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes (FStar.Seq.Base.append #quad32\n raw_quad_seq (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b))) 0 total_bytes in let\n (pb:(seq nat8)) = Vale.AES.GCTR_s.pad_to_128_bits ab in Vale.Def.Types_s.le_bytes_to_seq_quad32\n pb) (fun _ -> raw_quad_seq) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 460 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (auth_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32 raw_quad_seq\n (FStar.Seq.Base.create #quad32 1 length_quad) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 461 column 106 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_xmm 8 va_sM == Vale.AES.GCTR_s.gctr_encrypt_block ctr_BE_1 (Vale.AES.GHash_s.ghash_LE\n h_LE auth_quad_seq) alg key 0)))))))))))))))))) in\n assert_norm (va_qc.mods == va_mods);\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet\n 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13;\n va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm\n 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64\n rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64\n rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64\n rRax; va_Mod_ok; va_Mod_mem]) va_sM va_s0;\n (va_sM, va_fM)\n\n[@ va_qattr]\nlet va_wp_Gcm_blocks (alg:algorithm) (offset:int) (auth_b:buffer128) (abytes_b:buffer128)\n (in128x6_b:buffer128) (out128x6_b:buffer128) (in128_b:buffer128) (out128_b:buffer128)\n (inout_b:buffer128) (iv_b:buffer128) (scratch_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (va_s0:va_state) (va_k:(va_state -> unit ->\n Type0)) : Type0 =\n (va_get_ok va_s0 /\\ (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64\n (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0) in let\n (in128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 8) (va_get_stack va_s0) in let (out128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack va_s0) in\n let (len128x6:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 24) (va_get_stack va_s0) in let (in128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack va_s0) in\n let (out128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 40) (va_get_stack va_s0) in let (len128:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack va_s0) in\n let (inout_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 56) (va_get_stack va_s0) in let (plain_num_bytes:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack va_s0) in\n let (h_LE:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.X64.Decls.buffer128_read hkeys_b 2 (va_get_mem_heaplet 0 va_s0)) in sse_enabled /\\\n movbe_enabled /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 0)\n (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64\n (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 24) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 40) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 56) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet\n 1 va_s0) (va_get_reg64 rRdi va_s0) auth_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0)\n Secret /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 7 va_s0) abytes_ptr abytes_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 2\n va_s0) (va_get_reg64 rR8 va_s0) iv_b 1 (va_get_mem_layout va_s0) Public /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 6 va_s0) in128x6_ptr in128x6_b len128x6\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 6\n va_s0) out128x6_ptr out128x6_b len128x6 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0) in128_ptr in128_b len128\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1\n va_s0) out128_ptr out128_b len128 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) inout_ptr inout_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 3\n va_s0) (va_get_reg64 rRbp va_s0) scratch_b 9 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0) hkeys_b 8\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.buffer_disjoints128 iv_b ([keys_b;\n scratch_b; in128x6_b; out128x6_b; hkeys_b; in128_b; out128_b; inout_b]) /\\\n Vale.X64.Decls.buffer_disjoints128 scratch_b ([keys_b; in128x6_b; out128x6_b; in128_b;\n out128_b; inout_b; hkeys_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128x6_b ([keys_b;\n hkeys_b; in128_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128_b ([keys_b; hkeys_b;\n out128x6_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 inout_b ([keys_b; hkeys_b;\n out128x6_b; out128_b]) /\\ (Vale.X64.Decls.buffers_disjoint128 in128x6_b out128x6_b \\/ in128x6_b\n == out128x6_b) /\\ (Vale.X64.Decls.buffers_disjoint128 in128_b out128_b \\/ in128_b == out128_b)\n /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ in128x6_ptr\n + 16 `op_Multiply` len128x6 < pow2_64 /\\ out128x6_ptr + 16 `op_Multiply` len128x6 < pow2_64 /\\\n in128_ptr + 16 `op_Multiply` len128 < pow2_64 /\\ out128_ptr + 16 `op_Multiply` len128 < pow2_64\n /\\ inout_ptr + 16 < pow2_64 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 auth_b ==\n va_get_reg64 rRdx va_s0 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 abytes_b == 1\n /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128x6_b ==\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out128x6_b /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 out128_b /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128\n in128x6_b == len128x6 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b ==\n len128 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 inout_b == 1 /\\\n plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32 /\\ va_get_reg64 rR9 va_s0 + 32 <\n pow2_64 /\\ Vale.X64.Memory.buffer_addr #Vale.X64.Memory.vuint128 keys_b (va_get_mem_heaplet 0\n va_s0) + 128 < pow2_64 /\\ len128x6 `op_Modulus` 6 == 0 /\\ (len128x6 > 0 ==> len128x6 >= 6) /\\\n 12 + len128x6 + 6 < pow2_32 /\\ (va_mul_nat len128x6 (128 `op_Division` 8) + va_mul_nat len128\n (128 `op_Division` 8) <= plain_num_bytes /\\ plain_num_bytes < va_mul_nat len128x6 (128\n `op_Division` 8) + va_mul_nat len128 (128 `op_Division` 8) + 128 `op_Division` 8) /\\\n (va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) <= va_get_reg64 rRsi va_s0 /\\\n va_get_reg64 rRsi va_s0 < va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) + 128\n `op_Division` 8) /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rRcx va_s0)\n (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 (Vale.AES.AES_s.aes_encrypt_LE alg key\n (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0)))) /\\ (forall (va_x_mem:vale_heap)\n (va_x_rax:nat64) (va_x_rbx:nat64) (va_x_rcx:nat64) (va_x_rdx:nat64) (va_x_rdi:nat64)\n (va_x_rsi:nat64) (va_x_rbp:nat64) (va_x_r8:nat64) (va_x_r9:nat64) (va_x_r10:nat64)\n (va_x_r11:nat64) (va_x_r12:nat64) (va_x_r13:nat64) (va_x_r14:nat64) (va_x_r15:nat64)\n (va_x_xmm0:quad32) (va_x_xmm1:quad32) (va_x_xmm2:quad32) (va_x_xmm3:quad32) (va_x_xmm4:quad32)\n (va_x_xmm5:quad32) (va_x_xmm6:quad32) (va_x_xmm7:quad32) (va_x_xmm8:quad32) (va_x_xmm9:quad32)\n (va_x_xmm10:quad32) (va_x_xmm11:quad32) (va_x_xmm12:quad32) (va_x_xmm13:quad32)\n (va_x_xmm14:quad32) (va_x_xmm15:quad32) (va_x_heap1:vale_heap) (va_x_heap2:vale_heap)\n (va_x_heap3:vale_heap) (va_x_heap5:vale_heap) (va_x_heap6:vale_heap)\n (va_x_efl:Vale.X64.Flags.t) . let va_sM = va_upd_flags va_x_efl (va_upd_mem_heaplet 6\n va_x_heap6 (va_upd_mem_heaplet 5 va_x_heap5 (va_upd_mem_heaplet 3 va_x_heap3\n (va_upd_mem_heaplet 2 va_x_heap2 (va_upd_mem_heaplet 1 va_x_heap1 (va_upd_xmm 15 va_x_xmm15\n (va_upd_xmm 14 va_x_xmm14 (va_upd_xmm 13 va_x_xmm13 (va_upd_xmm 12 va_x_xmm12 (va_upd_xmm 11\n va_x_xmm11 (va_upd_xmm 10 va_x_xmm10 (va_upd_xmm 9 va_x_xmm9 (va_upd_xmm 8 va_x_xmm8\n (va_upd_xmm 7 va_x_xmm7 (va_upd_xmm 6 va_x_xmm6 (va_upd_xmm 5 va_x_xmm5 (va_upd_xmm 4 va_x_xmm4\n (va_upd_xmm 3 va_x_xmm3 (va_upd_xmm 2 va_x_xmm2 (va_upd_xmm 1 va_x_xmm1 (va_upd_xmm 0 va_x_xmm0\n (va_upd_reg64 rR15 va_x_r15 (va_upd_reg64 rR14 va_x_r14 (va_upd_reg64 rR13 va_x_r13\n (va_upd_reg64 rR12 va_x_r12 (va_upd_reg64 rR11 va_x_r11 (va_upd_reg64 rR10 va_x_r10\n (va_upd_reg64 rR9 va_x_r9 (va_upd_reg64 rR8 va_x_r8 (va_upd_reg64 rRbp va_x_rbp (va_upd_reg64\n rRsi va_x_rsi (va_upd_reg64 rRdi va_x_rdi (va_upd_reg64 rRdx va_x_rdx (va_upd_reg64 rRcx\n va_x_rcx (va_upd_reg64 rRbx va_x_rbx (va_upd_reg64 rRax va_x_rax (va_upd_mem va_x_mem\n va_s0))))))))))))))))))))))))))))))))))))) in va_get_ok va_sM /\\ (let\n (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in let (h_LE:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.X64.Decls.buffer128_read hkeys_b 2\n (va_get_mem_heaplet 0 va_s0)) in Vale.X64.Decls.modifies_buffer128 out128_b (va_get_mem_heaplet\n 1 va_s0) (va_get_mem_heaplet 1 va_sM) /\\ Vale.X64.Decls.modifies_buffer128 iv_b\n (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2 va_sM) /\\ Vale.X64.Decls.modifies_buffer128\n scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3 va_sM) /\\\n Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM) /\\ Vale.X64.Decls.modifies_buffer128 out128x6_b (va_get_mem_heaplet 6 va_s0)\n (va_get_mem_heaplet 6 va_sM) /\\ plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32\n /\\ (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in let\n (ctr_BE_1:quad32) = iv_BE in let (ctr_BE_2:quad32) = Vale.AES.GCTR_s.inc32 iv_BE 1 in let\n (plain_in:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b)) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in128_b)) in let (cipher_out:(seq quad32)) = va_if\n (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128 `op_Division` 8) (fun _ ->\n FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_sM) out128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM)\n inout_b)) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM)\n out128_b)) in let (cipher_bound:nat) = va_if (plain_num_bytes > (len128x6 + len128)\n `op_Multiply` 128 `op_Division` 8) (fun _ -> len128x6 + len128 + 1) (fun _ -> len128x6 +\n len128) in Vale.AES.GCTR.gctr_partial alg cipher_bound plain_in cipher_out key ctr_BE_2 /\\ (let\n (length_quad:quad32) = Vale.Def.Types_s.reverse_bytes_quad32 (Vale.Def.Types_s.insert_nat64\n (Vale.Def.Types_s.insert_nat64 (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0) (8\n `op_Multiply` va_get_reg64 rRsi va_s0) 1) (8 `op_Multiply` plain_num_bytes) 0) in let\n (raw_auth_quads:(seq quad32)) = va_if (va_get_reg64 rRsi va_s0 > va_get_reg64 rRdx va_s0\n `op_Multiply` 128 `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 7 va_s0) abytes_b)) (fun _ -> Vale.X64.Decls.s128 (va_get_mem_heaplet 1\n va_s0) auth_b) in let (auth_input_bytes:(seq nat8)) = FStar.Seq.Base.slice\n #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_auth_quads) 0 (va_get_reg64\n rRsi va_s0) in let (padded_auth_bytes:(seq nat8)) = Vale.AES.GCTR_s.pad_to_128_bits\n auth_input_bytes in let (auth_quad_seq:(seq quad32)) = Vale.Def.Types_s.le_bytes_to_seq_quad32\n padded_auth_bytes in let (raw_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32\n (FStar.Seq.Base.append #quad32 auth_quad_seq (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0)\n in128x6_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b) in let (total_bytes:nat)\n = FStar.Seq.Base.length #quad32 auth_quad_seq `op_Multiply` 16 + plain_num_bytes in let\n (raw_quad_seq:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> let (ab:(seq nat8)) = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes (FStar.Seq.Base.append #quad32 raw_quad_seq\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b))) 0 total_bytes in let (pb:(seq\n nat8)) = Vale.AES.GCTR_s.pad_to_128_bits ab in Vale.Def.Types_s.le_bytes_to_seq_quad32 pb) (fun\n _ -> raw_quad_seq) in let (auth_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32\n raw_quad_seq (FStar.Seq.Base.create #quad32 1 length_quad) in va_get_xmm 8 va_sM ==\n Vale.AES.GCTR_s.gctr_encrypt_block ctr_BE_1 (Vale.AES.GHash_s.ghash_LE h_LE auth_quad_seq) alg\n key 0))) ==> va_k va_sM (())))\n\n\nval va_wpProof_Gcm_blocks : alg:algorithm -> offset:int -> auth_b:buffer128 -> abytes_b:buffer128\n -> in128x6_b:buffer128 -> out128x6_b:buffer128 -> in128_b:buffer128 -> out128_b:buffer128 ->\n inout_b:buffer128 -> iv_b:buffer128 -> scratch_b:buffer128 -> key:(seq nat32) -> round_keys:(seq\n quad32) -> keys_b:buffer128 -> hkeys_b:buffer128 -> va_s0:va_state -> va_k:(va_state -> unit ->\n Type0)\n -> Ghost (va_state & va_fuel & unit)\n (requires (va_t_require va_s0 /\\ va_wp_Gcm_blocks alg offset auth_b abytes_b in128x6_b out128x6_b\n in128_b out128_b inout_b iv_b scratch_b key round_keys keys_b hkeys_b va_s0 va_k))\n (ensures (fun (va_sM, va_f0, va_g) -> va_t_ensure (va_code_Gcm_blocks alg offset) ([va_Mod_flags;\n va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet 3; va_Mod_mem_heaplet 2;\n va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13; va_Mod_xmm 12; va_Mod_xmm\n 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5;\n va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR15;\n va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64 rR11; va_Mod_reg64 rR10;\n va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64 rRsi; va_Mod_reg64 rRdi;\n va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64 rRax; va_Mod_mem]) va_s0\n va_k ((va_sM, va_f0, va_g))))\n[@\"opaque_to_smt\"]\nlet va_wpProof_Gcm_blocks alg offset auth_b abytes_b in128x6_b out128x6_b in128_b out128_b inout_b\n iv_b scratch_b key round_keys keys_b hkeys_b va_s0 va_k =\n let (va_sM, va_f0) = va_lemma_Gcm_blocks (va_code_Gcm_blocks alg offset) va_s0 alg offset auth_b\n abytes_b in128x6_b out128x6_b in128_b out128_b inout_b iv_b scratch_b key round_keys keys_b\n hkeys_b in\n va_lemma_upd_update va_sM;\n assert (va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 6 va_sM\n (va_update_mem_heaplet 5 va_sM (va_update_mem_heaplet 3 va_sM (va_update_mem_heaplet 2 va_sM\n (va_update_mem_heaplet 1 va_sM (va_update_xmm 15 va_sM (va_update_xmm 14 va_sM (va_update_xmm\n 13 va_sM (va_update_xmm 12 va_sM (va_update_xmm 11 va_sM (va_update_xmm 10 va_sM (va_update_xmm\n 9 va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5\n va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1\n va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR15 va_sM (va_update_reg64 rR14 va_sM\n (va_update_reg64 rR13 va_sM (va_update_reg64 rR12 va_sM (va_update_reg64 rR11 va_sM\n (va_update_reg64 rR10 va_sM (va_update_reg64 rR9 va_sM (va_update_reg64 rR8 va_sM\n (va_update_reg64 rRbp va_sM (va_update_reg64 rRsi va_sM (va_update_reg64 rRdi va_sM\n (va_update_reg64 rRdx va_sM (va_update_reg64 rRcx va_sM (va_update_reg64 rRbx va_sM\n (va_update_reg64 rRax va_sM (va_update_ok va_sM (va_update_mem va_sM\n va_s0))))))))))))))))))))))))))))))))))))))));\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet\n 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13;\n va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm\n 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64\n rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64\n rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64\n rRax; va_Mod_mem]) va_sM va_s0;\n let va_g = () in\n (va_sM, va_f0, va_g)\n\nPartial implementation:\nlet va_quick_Gcm_blocks\n (alg: algorithm)\n (offset: int)\n (auth_b abytes_b in128x6_b out128x6_b in128_b out128_b inout_b iv_b scratch_b: buffer128)\n (key: (seq nat32))\n (round_keys: (seq quad32))\n (keys_b hkeys_b: buffer128)\n : (va_quickCode unit (va_code_Gcm_blocks alg offset)) =\n \n"
},
{
"role": "assistant",
"content": "(va_QProc (va_code_Gcm_blocks alg offset)\n ([\n va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet 3;\n va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13;\n va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1;\n va_Mod_xmm 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12;\n va_Mod_reg64 rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp;\n va_Mod_reg64 rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx;\n va_Mod_reg64 rRbx; va_Mod_reg64 rRax; va_Mod_mem\n ])\n (va_wp_Gcm_blocks alg offset auth_b abytes_b in128x6_b out128x6_b in128_b out128_b inout_b iv_b\n scratch_b key round_keys keys_b hkeys_b)\n (va_wpProof_Gcm_blocks alg offset auth_b abytes_b in128x6_b out128x6_b in128_b out128_b inout_b\n iv_b scratch_b key round_keys keys_b hkeys_b))"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: index: Vale.Def.Types_s.nat8 -> dst: Vale.X64.Machine_s.quad32 -> src: Vale.X64.Machine_s.nat64\n -> FStar.Pervasives.Native.option Vale.X64.Machine_s.quad32\nEffect: Prims.Tot\n\nContext:\nmodule Vale.X64.Instructions_s\nopen FStar.Mul\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Two_s\nopen Vale.Def.Words.Four_s\nopen Vale.Def.Types_s\nopen Vale.X64.Machine_s\nopen Vale.X64.Instruction_s\nopen Vale.X64.CPU_Features_s\nopen Vale.X64.CryptoInstructions_s\nopen FStar.Seq.Base\n\nlet eval_Mov64 (src:nat64) : option nat64 = Some src\nval ins_Mov64 : instr_dep [out op64] [op64] PreserveFlags eval_Mov64\n\nlet eval_MovBe64 (src:nat64) : option nat64 = if movbe_enabled then Some (reverse_bytes_nat64 src) else None\nval ins_MovBe64 : instr_dep [out op64] [op64] PreserveFlags eval_MovBe64\n\nlet eval_Bswap64 (dst:nat64) : option nat64 = Some (reverse_bytes_nat64 dst)\nval ins_Bswap64 : instr_dep [inOut op64] [] PreserveFlags eval_Bswap64\n\nlet eval_Cmovc64 (dst src:nat64) (carry:bool) : option nat64 = Some (if carry then src else dst)\nval ins_Cmovc64 : instr_dep [inOut op64] [op64; opFlagsCf] PreserveFlags eval_Cmovc64\n\nlet eval_Add64 (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src in Some (sum >= pow2_64, sum % pow2_64)\nval ins_Add64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Add64\n\nlet eval_AddLea64 (src1 src2:nat64) : option nat64 = Some ((src1 + src2) % pow2_64)\nval ins_AddLea64 :instr_dep [out op64] [op64; op64] PreserveFlags eval_AddLea64\n\nlet eval_AddCarry64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_carry then 1 else 0) in\n Some (sum >= pow2_64, sum % pow2_64)\nval ins_AddCarry64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_AddCarry64\n\nlet eval_Adcx64_Adox64 (old_flag:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_flag then 1 else 0) in\n if adx_enabled then Some (sum >= pow2_64, sum % pow2_64) else None\nval ins_Adcx64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\nval ins_Adox64 : instr_dep [inOut opFlagsOf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\n\nlet eval_Sub64 (dst src:nat64) : option (bool & nat64) =\n let diff = dst - src in Some (diff < 0, diff % pow2_64)\nval ins_Sub64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Sub64\n\nlet eval_Sbb64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let diff = dst - (src + (if old_carry then 1 else 0)) in\n Some (diff < 0, diff % pow2_64) // We specify cf, but underspecify everything else (via HavocFlags)\nval ins_Sbb64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_Sbb64\n\nlet eval_Mul64 (rax src:nat64) : option (nat64 & nat64) =\n Some (FStar.UInt.mul_div #64 rax src, FStar.UInt.mul_mod #64 rax src)\nval ins_Mul64 : instr_dep [out (one64Reg rRdx); inOut (one64Reg rRax)] [op64] HavocFlags eval_Mul64\n\nlet eval_Mulx64 (rdx src:nat64) : option (nat64 & nat64) =\n let hi = FStar.UInt.mul_div #64 rdx src in\n let lo = FStar.UInt.mul_mod #64 rdx src in\n if bmi2_enabled then Some (hi, lo) else None\nval ins_Mulx64 : instr_dep [out op64; out op64] [one64Reg rRdx; op64] PreserveFlags eval_Mulx64\n\nlet eval_IMul64 (dst src:nat64) : option nat64 =\n Some (FStar.UInt.mul_mod #64 dst src)\nval ins_IMul64 : instr_dep [inOut op64] [op64] HavocFlags eval_IMul64\n\nlet eval_And64 (dst src:nat64) : option nat64 = Some (iand dst src)\nval ins_And64 : instr_dep [inOut op64] [op64] HavocFlags eval_And64\n\nlet eval_Xor64 (dst src:nat64) : option (nat64 & (bool & bool)) =\n Some (Vale.Def.Types_s.ixor dst src, (false, false))\nval ins_Xor64 : instr_dep [inOut op64; out opFlagsCf; out opFlagsOf] [op64] HavocFlags eval_Xor64\n\nlet eval_Shr64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishr dst amt) else None\nval ins_Shr64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shr64\n\nlet eval_Shl64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishl dst amt) else None\nval ins_Shl64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shl64\n\nlet eval_Cpuid (rax rcx:nat64) : option (nat64 & (nat64 & (nat64 & nat64))) =\n Some (cpuid rRax rax rcx, (cpuid rRbx rax rcx, (cpuid rRcx rax rcx, cpuid rRdx rax rcx)))\nval ins_Cpuid :\n instr_dep\n [inOut (one64Reg rRax); out (one64Reg rRbx); inOut (one64Reg rRcx); out (one64Reg rRdx)]\n [] PreserveFlags eval_Cpuid\n\n// The XGETBV instruction requires that OSXSAVE (in CPUID) is enabled.\n// We underspecify XGETBV here to only support fetching XCR0, which\n// is supported on any processor supporting the XGETBV instruction\nlet eval_Xgetbv (rcx:nat64) : option (nat64 & nat64) =\n if osxsave_enabled && rcx = 0 then Some (xgetbv rRax rcx, xgetbv rRdx rcx) else None\nval ins_Xgetbv :\n instr_dep\n [out (one64Reg rRax); out (one64Reg rRdx)]\n [one64Reg rRcx] PreserveFlags eval_Xgetbv\n\nlet check_avx (#a:Type0) (x:option a) : option a =\n if avx_enabled then x else None\n\nlet check_sse2 (#a:Type0) (x:option a) : option a =\n if sse2_enabled then x else None\n\nlet check_ssse3 (#a:Type0) (x:option a) : option a =\n if ssse3_enabled then x else None\n\nlet check_sse4_1 (#a:Type0) (x:option a) : option a =\n if sse4_1_enabled then x else None\n\nlet eval_Movdqu (src:quad32) : option quad32 = check_sse2 (Some src)\nval ins_Movdqu : instr_dep [out opXmm] [opXmm] PreserveFlags eval_Movdqu\n\nlet eval_Pxor (dst src:quad32) : option quad32 = check_sse2 (Some (quad32_xor dst src))\nval ins_Pxor : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pxor\n\nlet eval_VPxor (src1 src2:quad32) : option quad32 = check_avx (Some (quad32_xor src1 src2))\nval ins_VPxor : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPxor\n\nlet eval_Pand (dst src:quad32) : option quad32 = check_sse2 (Some (four_map2 (fun di si -> iand di si) dst src))\nval ins_Pand : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pand\n\nlet eval_Paddd_raw (src1 src2:quad32) : option quad32 =\n Some (Mkfour\n ((src1.lo0 + src2.lo0) % pow2_32)\n ((src1.lo1 + src2.lo1) % pow2_32)\n ((src1.hi2 + src2.hi2) % pow2_32)\n ((src1.hi3 + src2.hi3) % pow2_32))\n\nlet eval_Paddd (src1 src2:quad32) : option quad32 = check_sse2 (eval_Paddd_raw src1 src2)\nval ins_Paddd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Paddd\n\nlet eval_VPaddd (src1 src2:quad32) : option quad32 = check_avx (eval_Paddd_raw src1 src2)\nval ins_VPaddd : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPaddd\n\nlet eval_Pslld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishl i amt) dst) else None)\nval ins_Pslld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Pslld amt)\n\nlet eval_Psrld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishr i amt) dst) else None)\nval ins_Psrld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrld amt)\n\nlet eval_Psrldq (amt:int) (dst:quad32) : option quad32 = check_sse2 (\n if 0 <= amt && amt < 16 then\n let src_bytes = le_quad32_to_bytes dst in\n let zero_pad = Seq.create amt 0 in\n let remaining_bytes = slice src_bytes amt (length src_bytes) in\n Some (le_bytes_to_quad32 (append zero_pad remaining_bytes))\n else None)\nval ins_Psrldq (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrldq amt)\n\nlet eval_Palignr_raw (amount:nat8) (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if amount = 4 then\n Some (Mkfour src2.lo1 src2.hi2 src2.hi3 src1.lo0)\n else if amount = 8 then\n Some (Mkfour src2.hi2 src2.hi3 src1.lo0 src1.lo1)\n else None\n\nlet eval_Palignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_ssse3 (eval_Palignr_raw amount src1 src2)\nval ins_Palignr (amount:nat8) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Palignr amount)\n\nlet eval_VPalignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Palignr_raw amount src1 src2)\nval ins_VPalignr (amount:nat8) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPalignr amount)\n\nlet eval_Shufpd_raw (permutation:int) (src1 src2:quad32) : option quad32 =\n if 0 <= permutation && permutation < 4 then\n Some (Mkfour\n (if permutation % 2 = 0 then src1.lo0 else src1.hi2)\n (if permutation % 2 = 0 then src1.lo1 else src1.hi3)\n (if (permutation / 2) % 2 = 0 then src2.lo0 else src2.hi2)\n (if (permutation / 2) % 2 = 0 then src2.lo1 else src2.hi3))\n else None\n\nlet eval_Shufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_sse2 (eval_Shufpd_raw permutation src1 src2)\nval ins_Shufpd (permutation:int) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Shufpd permutation)\n\nlet eval_VShufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Shufpd_raw permutation src1 src2)\nval ins_VShufpd (permutation:int) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VShufpd permutation)\n\nlet is_full_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x04050607 &&\n q.hi3 = 0x00010203\n\nlet is_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x00010203 &&\n q.lo1 = 0x04050607 &&\n q.hi2 = 0x08090A0B &&\n q.hi3 = 0x0C0D0E0F\n\nlet is_high_dup_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet is_lower_upper_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x04050607 &&\n q.lo1 = 0x00010203 &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet eval_Pshufb_raw (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if is_full_byte_reversal_mask src2 then\n Some (reverse_bytes_quad32 src1)\n else if is_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3))\n else if is_high_dup_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else if is_lower_upper_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else None\n\nlet eval_Pshufb (src1 src2:quad32) : option quad32 = check_ssse3 (eval_Pshufb_raw src1 src2)\nval ins_Pshufb : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pshufb\n\nlet eval_VPshufb (src1 src2:quad32) : option quad32 = check_avx (eval_Pshufb_raw src1 src2)\nval ins_VPshufb : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPshufb\n\nlet eval_Pshufd (permutation:nat8) (src:quad32) : option quad32 = check_sse2 (\n let bits:bits_of_byte = byte_to_twobits permutation in\n Some (Mkfour\n (select_word src bits.lo0)\n (select_word src bits.lo1)\n (select_word src bits.hi2)\n (select_word src bits.hi3)))\nval ins_Pshufd (permutation:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_Pshufd permutation)\n\nlet eval_Pcmpeqd (dst src:quad32) : option quad32 = check_sse2 (\n let eq_result (b:bool):nat32 = if b then 0xFFFFFFFF else 0 in\n Some (Mkfour\n (eq_result (src.lo0 = dst.lo0))\n (eq_result (src.lo1 = dst.lo1))\n (eq_result (src.hi2 = dst.hi2))\n (eq_result (src.hi3 = dst.hi3))))\nval ins_Pcmpeqd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pcmpeqd\n\nlet eval_Pextrq (index:nat8) (src:quad32) : option nat64 = check_sse4_1 (\n let src_two = four_to_two_two src in\n Some (two_to_nat 32 (two_select src_two (index % 2))))\nval ins_Pextrq (index:nat8) : instr_dep [out op64] [opXmm] PreserveFlags (eval_Pextrq index)\n\nPartial implementation:\nlet eval_Pinsrd (index: nat8) (dst: quad32) (src: nat64) : option quad32 =\n \n"
},
{
"role": "assistant",
"content": "check_sse4_1 (Some (insert_nat32 dst (src % pow2_32) (index % 4)))"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: dst: Vale.X64.Machine_s.quad32 -> src: Vale.X64.Machine_s.quad32\n -> FStar.Pervasives.Native.option Vale.X64.Machine_s.quad32\nEffect: Prims.Tot\n\nContext:\nmodule Vale.X64.Instructions_s\nopen FStar.Mul\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Two_s\nopen Vale.Def.Words.Four_s\nopen Vale.Def.Types_s\nopen Vale.X64.Machine_s\nopen Vale.X64.Instruction_s\nopen Vale.X64.CPU_Features_s\nopen Vale.X64.CryptoInstructions_s\nopen FStar.Seq.Base\n\nlet eval_Mov64 (src:nat64) : option nat64 = Some src\nval ins_Mov64 : instr_dep [out op64] [op64] PreserveFlags eval_Mov64\n\nlet eval_MovBe64 (src:nat64) : option nat64 = if movbe_enabled then Some (reverse_bytes_nat64 src) else None\nval ins_MovBe64 : instr_dep [out op64] [op64] PreserveFlags eval_MovBe64\n\nlet eval_Bswap64 (dst:nat64) : option nat64 = Some (reverse_bytes_nat64 dst)\nval ins_Bswap64 : instr_dep [inOut op64] [] PreserveFlags eval_Bswap64\n\nlet eval_Cmovc64 (dst src:nat64) (carry:bool) : option nat64 = Some (if carry then src else dst)\nval ins_Cmovc64 : instr_dep [inOut op64] [op64; opFlagsCf] PreserveFlags eval_Cmovc64\n\nlet eval_Add64 (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src in Some (sum >= pow2_64, sum % pow2_64)\nval ins_Add64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Add64\n\nlet eval_AddLea64 (src1 src2:nat64) : option nat64 = Some ((src1 + src2) % pow2_64)\nval ins_AddLea64 :instr_dep [out op64] [op64; op64] PreserveFlags eval_AddLea64\n\nlet eval_AddCarry64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_carry then 1 else 0) in\n Some (sum >= pow2_64, sum % pow2_64)\nval ins_AddCarry64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_AddCarry64\n\nlet eval_Adcx64_Adox64 (old_flag:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_flag then 1 else 0) in\n if adx_enabled then Some (sum >= pow2_64, sum % pow2_64) else None\nval ins_Adcx64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\nval ins_Adox64 : instr_dep [inOut opFlagsOf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\n\nlet eval_Sub64 (dst src:nat64) : option (bool & nat64) =\n let diff = dst - src in Some (diff < 0, diff % pow2_64)\nval ins_Sub64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Sub64\n\nlet eval_Sbb64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let diff = dst - (src + (if old_carry then 1 else 0)) in\n Some (diff < 0, diff % pow2_64) // We specify cf, but underspecify everything else (via HavocFlags)\nval ins_Sbb64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_Sbb64\n\nlet eval_Mul64 (rax src:nat64) : option (nat64 & nat64) =\n Some (FStar.UInt.mul_div #64 rax src, FStar.UInt.mul_mod #64 rax src)\nval ins_Mul64 : instr_dep [out (one64Reg rRdx); inOut (one64Reg rRax)] [op64] HavocFlags eval_Mul64\n\nlet eval_Mulx64 (rdx src:nat64) : option (nat64 & nat64) =\n let hi = FStar.UInt.mul_div #64 rdx src in\n let lo = FStar.UInt.mul_mod #64 rdx src in\n if bmi2_enabled then Some (hi, lo) else None\nval ins_Mulx64 : instr_dep [out op64; out op64] [one64Reg rRdx; op64] PreserveFlags eval_Mulx64\n\nlet eval_IMul64 (dst src:nat64) : option nat64 =\n Some (FStar.UInt.mul_mod #64 dst src)\nval ins_IMul64 : instr_dep [inOut op64] [op64] HavocFlags eval_IMul64\n\nlet eval_And64 (dst src:nat64) : option nat64 = Some (iand dst src)\nval ins_And64 : instr_dep [inOut op64] [op64] HavocFlags eval_And64\n\nlet eval_Xor64 (dst src:nat64) : option (nat64 & (bool & bool)) =\n Some (Vale.Def.Types_s.ixor dst src, (false, false))\nval ins_Xor64 : instr_dep [inOut op64; out opFlagsCf; out opFlagsOf] [op64] HavocFlags eval_Xor64\n\nlet eval_Shr64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishr dst amt) else None\nval ins_Shr64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shr64\n\nlet eval_Shl64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishl dst amt) else None\nval ins_Shl64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shl64\n\nlet eval_Cpuid (rax rcx:nat64) : option (nat64 & (nat64 & (nat64 & nat64))) =\n Some (cpuid rRax rax rcx, (cpuid rRbx rax rcx, (cpuid rRcx rax rcx, cpuid rRdx rax rcx)))\nval ins_Cpuid :\n instr_dep\n [inOut (one64Reg rRax); out (one64Reg rRbx); inOut (one64Reg rRcx); out (one64Reg rRdx)]\n [] PreserveFlags eval_Cpuid\n\n// The XGETBV instruction requires that OSXSAVE (in CPUID) is enabled.\n// We underspecify XGETBV here to only support fetching XCR0, which\n// is supported on any processor supporting the XGETBV instruction\nlet eval_Xgetbv (rcx:nat64) : option (nat64 & nat64) =\n if osxsave_enabled && rcx = 0 then Some (xgetbv rRax rcx, xgetbv rRdx rcx) else None\nval ins_Xgetbv :\n instr_dep\n [out (one64Reg rRax); out (one64Reg rRdx)]\n [one64Reg rRcx] PreserveFlags eval_Xgetbv\n\nlet check_avx (#a:Type0) (x:option a) : option a =\n if avx_enabled then x else None\n\nlet check_sse2 (#a:Type0) (x:option a) : option a =\n if sse2_enabled then x else None\n\nlet check_ssse3 (#a:Type0) (x:option a) : option a =\n if ssse3_enabled then x else None\n\nlet check_sse4_1 (#a:Type0) (x:option a) : option a =\n if sse4_1_enabled then x else None\n\nlet eval_Movdqu (src:quad32) : option quad32 = check_sse2 (Some src)\nval ins_Movdqu : instr_dep [out opXmm] [opXmm] PreserveFlags eval_Movdqu\n\nlet eval_Pxor (dst src:quad32) : option quad32 = check_sse2 (Some (quad32_xor dst src))\nval ins_Pxor : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pxor\n\nlet eval_VPxor (src1 src2:quad32) : option quad32 = check_avx (Some (quad32_xor src1 src2))\nval ins_VPxor : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPxor\n\nlet eval_Pand (dst src:quad32) : option quad32 = check_sse2 (Some (four_map2 (fun di si -> iand di si) dst src))\nval ins_Pand : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pand\n\nlet eval_Paddd_raw (src1 src2:quad32) : option quad32 =\n Some (Mkfour\n ((src1.lo0 + src2.lo0) % pow2_32)\n ((src1.lo1 + src2.lo1) % pow2_32)\n ((src1.hi2 + src2.hi2) % pow2_32)\n ((src1.hi3 + src2.hi3) % pow2_32))\n\nlet eval_Paddd (src1 src2:quad32) : option quad32 = check_sse2 (eval_Paddd_raw src1 src2)\nval ins_Paddd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Paddd\n\nlet eval_VPaddd (src1 src2:quad32) : option quad32 = check_avx (eval_Paddd_raw src1 src2)\nval ins_VPaddd : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPaddd\n\nlet eval_Pslld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishl i amt) dst) else None)\nval ins_Pslld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Pslld amt)\n\nlet eval_Psrld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishr i amt) dst) else None)\nval ins_Psrld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrld amt)\n\nlet eval_Psrldq (amt:int) (dst:quad32) : option quad32 = check_sse2 (\n if 0 <= amt && amt < 16 then\n let src_bytes = le_quad32_to_bytes dst in\n let zero_pad = Seq.create amt 0 in\n let remaining_bytes = slice src_bytes amt (length src_bytes) in\n Some (le_bytes_to_quad32 (append zero_pad remaining_bytes))\n else None)\nval ins_Psrldq (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrldq amt)\n\nlet eval_Palignr_raw (amount:nat8) (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if amount = 4 then\n Some (Mkfour src2.lo1 src2.hi2 src2.hi3 src1.lo0)\n else if amount = 8 then\n Some (Mkfour src2.hi2 src2.hi3 src1.lo0 src1.lo1)\n else None\n\nlet eval_Palignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_ssse3 (eval_Palignr_raw amount src1 src2)\nval ins_Palignr (amount:nat8) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Palignr amount)\n\nlet eval_VPalignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Palignr_raw amount src1 src2)\nval ins_VPalignr (amount:nat8) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPalignr amount)\n\nlet eval_Shufpd_raw (permutation:int) (src1 src2:quad32) : option quad32 =\n if 0 <= permutation && permutation < 4 then\n Some (Mkfour\n (if permutation % 2 = 0 then src1.lo0 else src1.hi2)\n (if permutation % 2 = 0 then src1.lo1 else src1.hi3)\n (if (permutation / 2) % 2 = 0 then src2.lo0 else src2.hi2)\n (if (permutation / 2) % 2 = 0 then src2.lo1 else src2.hi3))\n else None\n\nlet eval_Shufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_sse2 (eval_Shufpd_raw permutation src1 src2)\nval ins_Shufpd (permutation:int) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Shufpd permutation)\n\nlet eval_VShufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Shufpd_raw permutation src1 src2)\nval ins_VShufpd (permutation:int) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VShufpd permutation)\n\nlet is_full_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x04050607 &&\n q.hi3 = 0x00010203\n\nlet is_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x00010203 &&\n q.lo1 = 0x04050607 &&\n q.hi2 = 0x08090A0B &&\n q.hi3 = 0x0C0D0E0F\n\nlet is_high_dup_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet is_lower_upper_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x04050607 &&\n q.lo1 = 0x00010203 &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet eval_Pshufb_raw (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if is_full_byte_reversal_mask src2 then\n Some (reverse_bytes_quad32 src1)\n else if is_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3))\n else if is_high_dup_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else if is_lower_upper_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else None\n\nlet eval_Pshufb (src1 src2:quad32) : option quad32 = check_ssse3 (eval_Pshufb_raw src1 src2)\nval ins_Pshufb : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pshufb\n\nlet eval_VPshufb (src1 src2:quad32) : option quad32 = check_avx (eval_Pshufb_raw src1 src2)\nval ins_VPshufb : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPshufb\n\nlet eval_Pshufd (permutation:nat8) (src:quad32) : option quad32 = check_sse2 (\n let bits:bits_of_byte = byte_to_twobits permutation in\n Some (Mkfour\n (select_word src bits.lo0)\n (select_word src bits.lo1)\n (select_word src bits.hi2)\n (select_word src bits.hi3)))\nval ins_Pshufd (permutation:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_Pshufd permutation)\n\nPartial implementation:\nlet eval_Pcmpeqd (dst src: quad32) : option quad32 =\n \n"
},
{
"role": "assistant",
"content": "check_sse2 (let eq_result (b: bool) : nat32 = if b then 0xFFFFFFFF else 0 in\n Some\n (Mkfour (eq_result (src.lo0 = dst.lo0))\n (eq_result (src.lo1 = dst.lo1))\n (eq_result (src.hi2 = dst.hi2))\n (eq_result (src.hi3 = dst.hi3))))"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: index: Vale.Def.Types_s.nat8 -> src: Vale.X64.Machine_s.quad32\n -> FStar.Pervasives.Native.option Vale.X64.Machine_s.nat64\nEffect: Prims.Tot\n\nContext:\nmodule Vale.X64.Instructions_s\nopen FStar.Mul\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Two_s\nopen Vale.Def.Words.Four_s\nopen Vale.Def.Types_s\nopen Vale.X64.Machine_s\nopen Vale.X64.Instruction_s\nopen Vale.X64.CPU_Features_s\nopen Vale.X64.CryptoInstructions_s\nopen FStar.Seq.Base\n\nlet eval_Mov64 (src:nat64) : option nat64 = Some src\nval ins_Mov64 : instr_dep [out op64] [op64] PreserveFlags eval_Mov64\n\nlet eval_MovBe64 (src:nat64) : option nat64 = if movbe_enabled then Some (reverse_bytes_nat64 src) else None\nval ins_MovBe64 : instr_dep [out op64] [op64] PreserveFlags eval_MovBe64\n\nlet eval_Bswap64 (dst:nat64) : option nat64 = Some (reverse_bytes_nat64 dst)\nval ins_Bswap64 : instr_dep [inOut op64] [] PreserveFlags eval_Bswap64\n\nlet eval_Cmovc64 (dst src:nat64) (carry:bool) : option nat64 = Some (if carry then src else dst)\nval ins_Cmovc64 : instr_dep [inOut op64] [op64; opFlagsCf] PreserveFlags eval_Cmovc64\n\nlet eval_Add64 (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src in Some (sum >= pow2_64, sum % pow2_64)\nval ins_Add64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Add64\n\nlet eval_AddLea64 (src1 src2:nat64) : option nat64 = Some ((src1 + src2) % pow2_64)\nval ins_AddLea64 :instr_dep [out op64] [op64; op64] PreserveFlags eval_AddLea64\n\nlet eval_AddCarry64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_carry then 1 else 0) in\n Some (sum >= pow2_64, sum % pow2_64)\nval ins_AddCarry64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_AddCarry64\n\nlet eval_Adcx64_Adox64 (old_flag:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_flag then 1 else 0) in\n if adx_enabled then Some (sum >= pow2_64, sum % pow2_64) else None\nval ins_Adcx64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\nval ins_Adox64 : instr_dep [inOut opFlagsOf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\n\nlet eval_Sub64 (dst src:nat64) : option (bool & nat64) =\n let diff = dst - src in Some (diff < 0, diff % pow2_64)\nval ins_Sub64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Sub64\n\nlet eval_Sbb64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let diff = dst - (src + (if old_carry then 1 else 0)) in\n Some (diff < 0, diff % pow2_64) // We specify cf, but underspecify everything else (via HavocFlags)\nval ins_Sbb64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_Sbb64\n\nlet eval_Mul64 (rax src:nat64) : option (nat64 & nat64) =\n Some (FStar.UInt.mul_div #64 rax src, FStar.UInt.mul_mod #64 rax src)\nval ins_Mul64 : instr_dep [out (one64Reg rRdx); inOut (one64Reg rRax)] [op64] HavocFlags eval_Mul64\n\nlet eval_Mulx64 (rdx src:nat64) : option (nat64 & nat64) =\n let hi = FStar.UInt.mul_div #64 rdx src in\n let lo = FStar.UInt.mul_mod #64 rdx src in\n if bmi2_enabled then Some (hi, lo) else None\nval ins_Mulx64 : instr_dep [out op64; out op64] [one64Reg rRdx; op64] PreserveFlags eval_Mulx64\n\nlet eval_IMul64 (dst src:nat64) : option nat64 =\n Some (FStar.UInt.mul_mod #64 dst src)\nval ins_IMul64 : instr_dep [inOut op64] [op64] HavocFlags eval_IMul64\n\nlet eval_And64 (dst src:nat64) : option nat64 = Some (iand dst src)\nval ins_And64 : instr_dep [inOut op64] [op64] HavocFlags eval_And64\n\nlet eval_Xor64 (dst src:nat64) : option (nat64 & (bool & bool)) =\n Some (Vale.Def.Types_s.ixor dst src, (false, false))\nval ins_Xor64 : instr_dep [inOut op64; out opFlagsCf; out opFlagsOf] [op64] HavocFlags eval_Xor64\n\nlet eval_Shr64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishr dst amt) else None\nval ins_Shr64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shr64\n\nlet eval_Shl64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishl dst amt) else None\nval ins_Shl64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shl64\n\nlet eval_Cpuid (rax rcx:nat64) : option (nat64 & (nat64 & (nat64 & nat64))) =\n Some (cpuid rRax rax rcx, (cpuid rRbx rax rcx, (cpuid rRcx rax rcx, cpuid rRdx rax rcx)))\nval ins_Cpuid :\n instr_dep\n [inOut (one64Reg rRax); out (one64Reg rRbx); inOut (one64Reg rRcx); out (one64Reg rRdx)]\n [] PreserveFlags eval_Cpuid\n\n// The XGETBV instruction requires that OSXSAVE (in CPUID) is enabled.\n// We underspecify XGETBV here to only support fetching XCR0, which\n// is supported on any processor supporting the XGETBV instruction\nlet eval_Xgetbv (rcx:nat64) : option (nat64 & nat64) =\n if osxsave_enabled && rcx = 0 then Some (xgetbv rRax rcx, xgetbv rRdx rcx) else None\nval ins_Xgetbv :\n instr_dep\n [out (one64Reg rRax); out (one64Reg rRdx)]\n [one64Reg rRcx] PreserveFlags eval_Xgetbv\n\nlet check_avx (#a:Type0) (x:option a) : option a =\n if avx_enabled then x else None\n\nlet check_sse2 (#a:Type0) (x:option a) : option a =\n if sse2_enabled then x else None\n\nlet check_ssse3 (#a:Type0) (x:option a) : option a =\n if ssse3_enabled then x else None\n\nlet check_sse4_1 (#a:Type0) (x:option a) : option a =\n if sse4_1_enabled then x else None\n\nlet eval_Movdqu (src:quad32) : option quad32 = check_sse2 (Some src)\nval ins_Movdqu : instr_dep [out opXmm] [opXmm] PreserveFlags eval_Movdqu\n\nlet eval_Pxor (dst src:quad32) : option quad32 = check_sse2 (Some (quad32_xor dst src))\nval ins_Pxor : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pxor\n\nlet eval_VPxor (src1 src2:quad32) : option quad32 = check_avx (Some (quad32_xor src1 src2))\nval ins_VPxor : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPxor\n\nlet eval_Pand (dst src:quad32) : option quad32 = check_sse2 (Some (four_map2 (fun di si -> iand di si) dst src))\nval ins_Pand : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pand\n\nlet eval_Paddd_raw (src1 src2:quad32) : option quad32 =\n Some (Mkfour\n ((src1.lo0 + src2.lo0) % pow2_32)\n ((src1.lo1 + src2.lo1) % pow2_32)\n ((src1.hi2 + src2.hi2) % pow2_32)\n ((src1.hi3 + src2.hi3) % pow2_32))\n\nlet eval_Paddd (src1 src2:quad32) : option quad32 = check_sse2 (eval_Paddd_raw src1 src2)\nval ins_Paddd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Paddd\n\nlet eval_VPaddd (src1 src2:quad32) : option quad32 = check_avx (eval_Paddd_raw src1 src2)\nval ins_VPaddd : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPaddd\n\nlet eval_Pslld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishl i amt) dst) else None)\nval ins_Pslld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Pslld amt)\n\nlet eval_Psrld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishr i amt) dst) else None)\nval ins_Psrld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrld amt)\n\nlet eval_Psrldq (amt:int) (dst:quad32) : option quad32 = check_sse2 (\n if 0 <= amt && amt < 16 then\n let src_bytes = le_quad32_to_bytes dst in\n let zero_pad = Seq.create amt 0 in\n let remaining_bytes = slice src_bytes amt (length src_bytes) in\n Some (le_bytes_to_quad32 (append zero_pad remaining_bytes))\n else None)\nval ins_Psrldq (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrldq amt)\n\nlet eval_Palignr_raw (amount:nat8) (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if amount = 4 then\n Some (Mkfour src2.lo1 src2.hi2 src2.hi3 src1.lo0)\n else if amount = 8 then\n Some (Mkfour src2.hi2 src2.hi3 src1.lo0 src1.lo1)\n else None\n\nlet eval_Palignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_ssse3 (eval_Palignr_raw amount src1 src2)\nval ins_Palignr (amount:nat8) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Palignr amount)\n\nlet eval_VPalignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Palignr_raw amount src1 src2)\nval ins_VPalignr (amount:nat8) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPalignr amount)\n\nlet eval_Shufpd_raw (permutation:int) (src1 src2:quad32) : option quad32 =\n if 0 <= permutation && permutation < 4 then\n Some (Mkfour\n (if permutation % 2 = 0 then src1.lo0 else src1.hi2)\n (if permutation % 2 = 0 then src1.lo1 else src1.hi3)\n (if (permutation / 2) % 2 = 0 then src2.lo0 else src2.hi2)\n (if (permutation / 2) % 2 = 0 then src2.lo1 else src2.hi3))\n else None\n\nlet eval_Shufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_sse2 (eval_Shufpd_raw permutation src1 src2)\nval ins_Shufpd (permutation:int) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Shufpd permutation)\n\nlet eval_VShufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Shufpd_raw permutation src1 src2)\nval ins_VShufpd (permutation:int) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VShufpd permutation)\n\nlet is_full_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x04050607 &&\n q.hi3 = 0x00010203\n\nlet is_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x00010203 &&\n q.lo1 = 0x04050607 &&\n q.hi2 = 0x08090A0B &&\n q.hi3 = 0x0C0D0E0F\n\nlet is_high_dup_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet is_lower_upper_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x04050607 &&\n q.lo1 = 0x00010203 &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet eval_Pshufb_raw (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if is_full_byte_reversal_mask src2 then\n Some (reverse_bytes_quad32 src1)\n else if is_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3))\n else if is_high_dup_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else if is_lower_upper_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else None\n\nlet eval_Pshufb (src1 src2:quad32) : option quad32 = check_ssse3 (eval_Pshufb_raw src1 src2)\nval ins_Pshufb : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pshufb\n\nlet eval_VPshufb (src1 src2:quad32) : option quad32 = check_avx (eval_Pshufb_raw src1 src2)\nval ins_VPshufb : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPshufb\n\nlet eval_Pshufd (permutation:nat8) (src:quad32) : option quad32 = check_sse2 (\n let bits:bits_of_byte = byte_to_twobits permutation in\n Some (Mkfour\n (select_word src bits.lo0)\n (select_word src bits.lo1)\n (select_word src bits.hi2)\n (select_word src bits.hi3)))\nval ins_Pshufd (permutation:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_Pshufd permutation)\n\nlet eval_Pcmpeqd (dst src:quad32) : option quad32 = check_sse2 (\n let eq_result (b:bool):nat32 = if b then 0xFFFFFFFF else 0 in\n Some (Mkfour\n (eq_result (src.lo0 = dst.lo0))\n (eq_result (src.lo1 = dst.lo1))\n (eq_result (src.hi2 = dst.hi2))\n (eq_result (src.hi3 = dst.hi3))))\nval ins_Pcmpeqd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pcmpeqd\n\nPartial implementation:\nlet eval_Pextrq (index: nat8) (src: quad32) : option nat64 =\n \n"
},
{
"role": "assistant",
"content": "check_sse4_1 (let src_two = four_to_two_two src in\n Some (two_to_nat 32 (two_select src_two (index % 2))))"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: index: Vale.Def.Types_s.nat8 -> dst: Vale.X64.Machine_s.quad32 -> src: Vale.X64.Machine_s.nat64\n -> FStar.Pervasives.Native.option Vale.X64.Machine_s.quad32\nEffect: Prims.Tot\n\nContext:\nmodule Vale.X64.Instructions_s\nopen FStar.Mul\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Two_s\nopen Vale.Def.Words.Four_s\nopen Vale.Def.Types_s\nopen Vale.X64.Machine_s\nopen Vale.X64.Instruction_s\nopen Vale.X64.CPU_Features_s\nopen Vale.X64.CryptoInstructions_s\nopen FStar.Seq.Base\n\nlet eval_Mov64 (src:nat64) : option nat64 = Some src\nval ins_Mov64 : instr_dep [out op64] [op64] PreserveFlags eval_Mov64\n\nlet eval_MovBe64 (src:nat64) : option nat64 = if movbe_enabled then Some (reverse_bytes_nat64 src) else None\nval ins_MovBe64 : instr_dep [out op64] [op64] PreserveFlags eval_MovBe64\n\nlet eval_Bswap64 (dst:nat64) : option nat64 = Some (reverse_bytes_nat64 dst)\nval ins_Bswap64 : instr_dep [inOut op64] [] PreserveFlags eval_Bswap64\n\nlet eval_Cmovc64 (dst src:nat64) (carry:bool) : option nat64 = Some (if carry then src else dst)\nval ins_Cmovc64 : instr_dep [inOut op64] [op64; opFlagsCf] PreserveFlags eval_Cmovc64\n\nlet eval_Add64 (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src in Some (sum >= pow2_64, sum % pow2_64)\nval ins_Add64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Add64\n\nlet eval_AddLea64 (src1 src2:nat64) : option nat64 = Some ((src1 + src2) % pow2_64)\nval ins_AddLea64 :instr_dep [out op64] [op64; op64] PreserveFlags eval_AddLea64\n\nlet eval_AddCarry64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_carry then 1 else 0) in\n Some (sum >= pow2_64, sum % pow2_64)\nval ins_AddCarry64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_AddCarry64\n\nlet eval_Adcx64_Adox64 (old_flag:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_flag then 1 else 0) in\n if adx_enabled then Some (sum >= pow2_64, sum % pow2_64) else None\nval ins_Adcx64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\nval ins_Adox64 : instr_dep [inOut opFlagsOf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\n\nlet eval_Sub64 (dst src:nat64) : option (bool & nat64) =\n let diff = dst - src in Some (diff < 0, diff % pow2_64)\nval ins_Sub64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Sub64\n\nlet eval_Sbb64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let diff = dst - (src + (if old_carry then 1 else 0)) in\n Some (diff < 0, diff % pow2_64) // We specify cf, but underspecify everything else (via HavocFlags)\nval ins_Sbb64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_Sbb64\n\nlet eval_Mul64 (rax src:nat64) : option (nat64 & nat64) =\n Some (FStar.UInt.mul_div #64 rax src, FStar.UInt.mul_mod #64 rax src)\nval ins_Mul64 : instr_dep [out (one64Reg rRdx); inOut (one64Reg rRax)] [op64] HavocFlags eval_Mul64\n\nlet eval_Mulx64 (rdx src:nat64) : option (nat64 & nat64) =\n let hi = FStar.UInt.mul_div #64 rdx src in\n let lo = FStar.UInt.mul_mod #64 rdx src in\n if bmi2_enabled then Some (hi, lo) else None\nval ins_Mulx64 : instr_dep [out op64; out op64] [one64Reg rRdx; op64] PreserveFlags eval_Mulx64\n\nlet eval_IMul64 (dst src:nat64) : option nat64 =\n Some (FStar.UInt.mul_mod #64 dst src)\nval ins_IMul64 : instr_dep [inOut op64] [op64] HavocFlags eval_IMul64\n\nlet eval_And64 (dst src:nat64) : option nat64 = Some (iand dst src)\nval ins_And64 : instr_dep [inOut op64] [op64] HavocFlags eval_And64\n\nlet eval_Xor64 (dst src:nat64) : option (nat64 & (bool & bool)) =\n Some (Vale.Def.Types_s.ixor dst src, (false, false))\nval ins_Xor64 : instr_dep [inOut op64; out opFlagsCf; out opFlagsOf] [op64] HavocFlags eval_Xor64\n\nlet eval_Shr64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishr dst amt) else None\nval ins_Shr64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shr64\n\nlet eval_Shl64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishl dst amt) else None\nval ins_Shl64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shl64\n\nlet eval_Cpuid (rax rcx:nat64) : option (nat64 & (nat64 & (nat64 & nat64))) =\n Some (cpuid rRax rax rcx, (cpuid rRbx rax rcx, (cpuid rRcx rax rcx, cpuid rRdx rax rcx)))\nval ins_Cpuid :\n instr_dep\n [inOut (one64Reg rRax); out (one64Reg rRbx); inOut (one64Reg rRcx); out (one64Reg rRdx)]\n [] PreserveFlags eval_Cpuid\n\n// The XGETBV instruction requires that OSXSAVE (in CPUID) is enabled.\n// We underspecify XGETBV here to only support fetching XCR0, which\n// is supported on any processor supporting the XGETBV instruction\nlet eval_Xgetbv (rcx:nat64) : option (nat64 & nat64) =\n if osxsave_enabled && rcx = 0 then Some (xgetbv rRax rcx, xgetbv rRdx rcx) else None\nval ins_Xgetbv :\n instr_dep\n [out (one64Reg rRax); out (one64Reg rRdx)]\n [one64Reg rRcx] PreserveFlags eval_Xgetbv\n\nlet check_avx (#a:Type0) (x:option a) : option a =\n if avx_enabled then x else None\n\nlet check_sse2 (#a:Type0) (x:option a) : option a =\n if sse2_enabled then x else None\n\nlet check_ssse3 (#a:Type0) (x:option a) : option a =\n if ssse3_enabled then x else None\n\nlet check_sse4_1 (#a:Type0) (x:option a) : option a =\n if sse4_1_enabled then x else None\n\nlet eval_Movdqu (src:quad32) : option quad32 = check_sse2 (Some src)\nval ins_Movdqu : instr_dep [out opXmm] [opXmm] PreserveFlags eval_Movdqu\n\nlet eval_Pxor (dst src:quad32) : option quad32 = check_sse2 (Some (quad32_xor dst src))\nval ins_Pxor : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pxor\n\nlet eval_VPxor (src1 src2:quad32) : option quad32 = check_avx (Some (quad32_xor src1 src2))\nval ins_VPxor : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPxor\n\nlet eval_Pand (dst src:quad32) : option quad32 = check_sse2 (Some (four_map2 (fun di si -> iand di si) dst src))\nval ins_Pand : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pand\n\nlet eval_Paddd_raw (src1 src2:quad32) : option quad32 =\n Some (Mkfour\n ((src1.lo0 + src2.lo0) % pow2_32)\n ((src1.lo1 + src2.lo1) % pow2_32)\n ((src1.hi2 + src2.hi2) % pow2_32)\n ((src1.hi3 + src2.hi3) % pow2_32))\n\nlet eval_Paddd (src1 src2:quad32) : option quad32 = check_sse2 (eval_Paddd_raw src1 src2)\nval ins_Paddd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Paddd\n\nlet eval_VPaddd (src1 src2:quad32) : option quad32 = check_avx (eval_Paddd_raw src1 src2)\nval ins_VPaddd : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPaddd\n\nlet eval_Pslld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishl i amt) dst) else None)\nval ins_Pslld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Pslld amt)\n\nlet eval_Psrld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishr i amt) dst) else None)\nval ins_Psrld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrld amt)\n\nlet eval_Psrldq (amt:int) (dst:quad32) : option quad32 = check_sse2 (\n if 0 <= amt && amt < 16 then\n let src_bytes = le_quad32_to_bytes dst in\n let zero_pad = Seq.create amt 0 in\n let remaining_bytes = slice src_bytes amt (length src_bytes) in\n Some (le_bytes_to_quad32 (append zero_pad remaining_bytes))\n else None)\nval ins_Psrldq (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrldq amt)\n\nlet eval_Palignr_raw (amount:nat8) (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if amount = 4 then\n Some (Mkfour src2.lo1 src2.hi2 src2.hi3 src1.lo0)\n else if amount = 8 then\n Some (Mkfour src2.hi2 src2.hi3 src1.lo0 src1.lo1)\n else None\n\nlet eval_Palignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_ssse3 (eval_Palignr_raw amount src1 src2)\nval ins_Palignr (amount:nat8) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Palignr amount)\n\nlet eval_VPalignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Palignr_raw amount src1 src2)\nval ins_VPalignr (amount:nat8) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPalignr amount)\n\nlet eval_Shufpd_raw (permutation:int) (src1 src2:quad32) : option quad32 =\n if 0 <= permutation && permutation < 4 then\n Some (Mkfour\n (if permutation % 2 = 0 then src1.lo0 else src1.hi2)\n (if permutation % 2 = 0 then src1.lo1 else src1.hi3)\n (if (permutation / 2) % 2 = 0 then src2.lo0 else src2.hi2)\n (if (permutation / 2) % 2 = 0 then src2.lo1 else src2.hi3))\n else None\n\nlet eval_Shufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_sse2 (eval_Shufpd_raw permutation src1 src2)\nval ins_Shufpd (permutation:int) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Shufpd permutation)\n\nlet eval_VShufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Shufpd_raw permutation src1 src2)\nval ins_VShufpd (permutation:int) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VShufpd permutation)\n\nlet is_full_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x04050607 &&\n q.hi3 = 0x00010203\n\nlet is_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x00010203 &&\n q.lo1 = 0x04050607 &&\n q.hi2 = 0x08090A0B &&\n q.hi3 = 0x0C0D0E0F\n\nlet is_high_dup_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet is_lower_upper_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x04050607 &&\n q.lo1 = 0x00010203 &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet eval_Pshufb_raw (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if is_full_byte_reversal_mask src2 then\n Some (reverse_bytes_quad32 src1)\n else if is_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3))\n else if is_high_dup_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else if is_lower_upper_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else None\n\nlet eval_Pshufb (src1 src2:quad32) : option quad32 = check_ssse3 (eval_Pshufb_raw src1 src2)\nval ins_Pshufb : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pshufb\n\nlet eval_VPshufb (src1 src2:quad32) : option quad32 = check_avx (eval_Pshufb_raw src1 src2)\nval ins_VPshufb : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPshufb\n\nlet eval_Pshufd (permutation:nat8) (src:quad32) : option quad32 = check_sse2 (\n let bits:bits_of_byte = byte_to_twobits permutation in\n Some (Mkfour\n (select_word src bits.lo0)\n (select_word src bits.lo1)\n (select_word src bits.hi2)\n (select_word src bits.hi3)))\nval ins_Pshufd (permutation:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_Pshufd permutation)\n\nlet eval_Pcmpeqd (dst src:quad32) : option quad32 = check_sse2 (\n let eq_result (b:bool):nat32 = if b then 0xFFFFFFFF else 0 in\n Some (Mkfour\n (eq_result (src.lo0 = dst.lo0))\n (eq_result (src.lo1 = dst.lo1))\n (eq_result (src.hi2 = dst.hi2))\n (eq_result (src.hi3 = dst.hi3))))\nval ins_Pcmpeqd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pcmpeqd\n\nlet eval_Pextrq (index:nat8) (src:quad32) : option nat64 = check_sse4_1 (\n let src_two = four_to_two_two src in\n Some (two_to_nat 32 (two_select src_two (index % 2))))\nval ins_Pextrq (index:nat8) : instr_dep [out op64] [opXmm] PreserveFlags (eval_Pextrq index)\n\nlet eval_Pinsrd (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat32 dst (src % pow2_32) (index % 4)))\nval ins_Pinsrd (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrd index)\n\nPartial implementation:\nlet eval_Pinsrq (index: nat8) (dst: quad32) (src: nat64) : option quad32 =\n \n"
},
{
"role": "assistant",
"content": "check_sse4_1 (Some (insert_nat64_def dst src (index % 2)))"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n ke: Hacl.Bignum.Exponentiation.exp t ->\n a: Spec.Hash.Definitions.hash_alg{Spec.RSAPSS.hash_is_supported a} ->\n modBits: Hacl.Impl.RSAPSS.modBits_t t\n -> Hacl.Impl.RSAPSS.rsapss_verify_st1 t ke a modBits\nEffect: Prims.Tot\n\nContext:\nmodule Hacl.Impl.RSAPSS\n\nopen FStar.HyperStack\nopen FStar.HyperStack.ST\nopen FStar.Mul\n\nopen Lib.IntTypes\nopen Lib.Buffer\n\nopen Hacl.Bignum.Definitions\n\nmodule ST = FStar.HyperStack.ST\n\nmodule Hash = Spec.Agile.Hash\nmodule SB = Hacl.Spec.Bignum\nmodule BB = Hacl.Spec.Bignum.Base\nmodule SD = Hacl.Spec.Bignum.Definitions\nmodule SM = Hacl.Spec.Bignum.Montgomery\nmodule SE = Hacl.Spec.Bignum.Exponentiation\n\nmodule BN = Hacl.Bignum\nmodule BE = Hacl.Bignum.Exponentiation\nmodule BM = Hacl.Bignum.Montgomery\n\nmodule S = Spec.RSAPSS\nmodule LS = Hacl.Spec.RSAPSS\nmodule LSeq = Lib.Sequence\n\nmodule RP = Hacl.Impl.RSAPSS.Padding\nmodule RM = Hacl.Impl.RSAPSS.MGF\nmodule RK = Hacl.Impl.RSAPSS.Keys\n\n#reset-options \"--z3rlimit 150 --fuel 0 --ifuel 0\"\n\ninline_for_extraction noextract\nlet modBits_t (t:limb_t) = modBits:size_t{1 < v modBits /\\ 2 * bits t * SD.blocks (v modBits) (bits t) <= max_size_t}\n\n\ninline_for_extraction noextract\nlet rsapss_sign_bn_st (t:limb_t) (ke:BE.exp t) (modBits:modBits_t t) =\n let len = blocks modBits (size (bits t)) in\n eBits:size_t\n -> dBits:size_t{LS.skey_len_pre t (v modBits) (v eBits) (v dBits)}\n -> skey:lbignum t (2ul *! len +! blocks eBits (size (bits t)) +! blocks dBits (size (bits t)))\n -> m:lbignum t len\n -> m':lbignum t len\n -> s:lbignum t len ->\n Stack bool\n (requires fun h -> len == ke.BE.bn.BN.len /\\\n live h skey /\\ live h m /\\ live h s /\\ live h m' /\\\n disjoint s m /\\ disjoint s skey /\\ disjoint m skey /\\\n disjoint m m' /\\ disjoint m' s /\\ disjoint m' skey /\\\n LS.rsapss_skey_pre (v modBits) (v eBits) (v dBits) (as_seq h skey) /\\\n bn_v h m < bn_v h (gsub skey 0ul len))\n (ensures fun h0 r h1 -> modifies (loc s |+| loc m') h0 h1 /\\\n (r, as_seq h1 s) == LS.rsapss_sign_bn (v modBits) (v eBits) (v dBits) (as_seq h0 skey) (as_seq h0 m))\n\n\ninline_for_extraction noextract\nval rsapss_sign_bn: #t:limb_t -> ke:BE.exp t -> modBits:modBits_t t -> rsapss_sign_bn_st t ke modBits\nlet rsapss_sign_bn #t ke modBits eBits dBits skey m m' s =\n [@inline_let] let bits : size_pos = bits t in\n let nLen = blocks modBits (size bits) in\n let eLen = blocks eBits (size bits) in\n let dLen = blocks dBits (size bits) in\n\n let n = sub skey 0ul nLen in\n let r2 = sub skey nLen nLen in\n let e = sub skey (nLen +! nLen) eLen in\n let d = sub skey (nLen +! nLen +! eLen) dLen in\n\n Math.Lemmas.pow2_le_compat (bits * v nLen) (v modBits);\n let h0 = ST.get () in\n SM.bn_precomp_r2_mod_n_lemma (v modBits - 1) (as_seq h0 n);\n BE.mk_bn_mod_exp_precompr2 nLen ke.BE.exp_ct_precomp n r2 m dBits d s;\n BE.mk_bn_mod_exp_precompr2 nLen ke.BE.exp_vt_precomp n r2 s eBits e m';\n let h1 = ST.get () in\n SD.bn_eval_inj (v nLen) (as_seq h1 s)\n (SE.bn_mod_exp_consttime_precompr2 (v nLen) (as_seq h0 n) (as_seq h0 r2)\n (as_seq h0 m) (v dBits) (as_seq h0 d));\n SD.bn_eval_inj (v nLen) (as_seq h1 m')\n (SE.bn_mod_exp_vartime_precompr2 (v nLen) (as_seq h0 n) (as_seq h0 r2)\n (as_seq h1 s) (v eBits) (as_seq h0 e));\n let eq_m = BN.bn_eq_mask nLen m m' in\n mapT nLen s (logand eq_m) s;\n BB.unsafe_bool_of_limb eq_m\n\n\ninline_for_extraction noextract\nlet rsapss_sign_msg_to_bn_st (t:limb_t) (a:Hash.hash_alg{S.hash_is_supported a}) (modBits:modBits_t t) =\n let len = blocks modBits (size (bits t)) in\n saltLen:size_t\n -> salt:lbuffer uint8 saltLen\n -> msgLen:size_t\n -> msg:lbuffer uint8 msgLen\n -> m:lbignum t len ->\n Stack unit\n (requires fun h ->\n live h salt /\\ live h msg /\\ live h m /\\\n disjoint salt msg /\\ disjoint m msg /\\ disjoint m salt /\\\n as_seq h m == LSeq.create (v len) (uint #t 0) /\\\n LS.rsapss_sign_pre a (v modBits) (v saltLen) (as_seq h salt) (v msgLen) (as_seq h msg))\n (ensures fun h0 _ h1 -> modifies (loc m) h0 h1 /\\\n as_seq h1 m == LS.rsapss_sign_msg_to_bn a (v modBits) (v saltLen) (as_seq h0 salt) (v msgLen) (as_seq h0 msg))\n\n\ninline_for_extraction noextract\nval rsapss_sign_msg_to_bn:\n #t:limb_t\n -> a:Hash.hash_alg{S.hash_is_supported a}\n -> modBits:modBits_t t ->\n rsapss_sign_msg_to_bn_st t a modBits\n\nlet rsapss_sign_msg_to_bn #t a modBits saltLen salt msgLen msg m =\n push_frame ();\n [@inline_let] let bits : size_pos = bits t in\n [@inline_let] let numb : size_pos = numbytes t in\n let nLen = blocks modBits (size bits) in\n\n let emBits = modBits -! 1ul in\n let emLen = blocks emBits 8ul in\n [@inline_let] let mLen = blocks emLen (size numb) in\n\n let em = create emLen (u8 0) in\n RP.pss_encode a saltLen salt msgLen msg emBits em;\n LS.blocks_bits_lemma t (v emBits);\n LS.blocks_numb_lemma t (v emBits);\n assert (SD.blocks (v emBits) bits = v mLen);\n assert (numb * v mLen <= max_size_t);\n assert (v mLen <= v nLen);\n\n let h' = ST.get () in\n update_sub_f h' m 0ul mLen\n (fun h -> SB.bn_from_bytes_be (v emLen) (as_seq h' em))\n (fun _ -> BN.bn_from_bytes_be emLen em (sub m 0ul mLen));\n pop_frame ()\n\n\ninline_for_extraction noextract\nlet rsapss_sign_compute_sgnt_st (t:limb_t) (ke:BE.exp t) (modBits:modBits_t t) =\n let len = blocks modBits (size (bits t)) in\n eBits:size_t\n -> dBits:size_t{LS.skey_len_pre t (v modBits) (v eBits) (v dBits)}\n -> skey:lbignum t (2ul *! len +! blocks eBits (size (bits t)) +! blocks dBits (size (bits t)))\n -> m:lbignum t len\n -> sgnt:lbuffer uint8 (blocks modBits 8ul) ->\n Stack bool\n (requires fun h -> len == ke.BE.bn.BN.len /\\\n live h sgnt /\\ live h skey /\\ live h m /\\\n disjoint sgnt skey /\\ disjoint m sgnt /\\ disjoint m skey /\\\n LS.rsapss_skey_pre (v modBits) (v eBits) (v dBits) (as_seq h skey) /\\\n bn_v h m < bn_v h (gsub skey 0ul len))\n (ensures fun h0 eq_m h1 -> modifies (loc sgnt) h0 h1 /\\\n (eq_m, as_seq h1 sgnt) == LS.rsapss_sign_compute_sgnt (v modBits) (v eBits) (v dBits) (as_seq h0 skey) (as_seq h0 m))\n\n\ninline_for_extraction noextract\nval rsapss_sign_compute_sgnt:\n #t:limb_t\n -> ke:BE.exp t\n -> modBits:modBits_t t ->\n rsapss_sign_compute_sgnt_st t ke modBits\n\nlet rsapss_sign_compute_sgnt #t ke modBits eBits dBits skey m sgnt =\n push_frame ();\n let h_init = ST.get () in\n [@inline_let] let bits : size_pos = bits t in\n [@inline_let] let numb : size_pos = numbytes t in\n let nLen = blocks modBits (size bits) in\n\n let k = blocks modBits 8ul in\n let s = create nLen (uint #t 0) in\n let m' = create nLen (uint #t 0) in\n let eq_b = rsapss_sign_bn ke modBits eBits dBits skey m m' s in\n LS.blocks_bits_lemma t (v modBits);\n LS.blocks_numb_lemma t (v modBits);\n assert (SD.blocks (v k) numb == v nLen);\n assert (numb * v nLen <= max_size_t);\n BN.bn_to_bytes_be k s sgnt;\n pop_frame ();\n eq_b\n\n\ninline_for_extraction noextract\nlet rsapss_sign_st1 (t:limb_t) (ke:BE.exp t) (a:Hash.hash_alg{S.hash_is_supported a}) (modBits:modBits_t t) =\n let len = blocks modBits (size (bits t)) in\n eBits:size_t\n -> dBits:size_t{LS.skey_len_pre t (v modBits) (v eBits) (v dBits)}\n -> skey:lbignum t (2ul *! len +! blocks eBits (size (bits t)) +! blocks dBits (size (bits t)))\n -> saltLen:size_t\n -> salt:lbuffer uint8 saltLen\n -> msgLen:size_t\n -> msg:lbuffer uint8 msgLen\n -> sgnt:lbuffer uint8 (blocks modBits 8ul) ->\n Stack bool\n (requires fun h -> len == ke.BE.bn.BN.len /\\\n live h salt /\\ live h msg /\\ live h sgnt /\\ live h skey /\\\n disjoint sgnt salt /\\ disjoint sgnt msg /\\ disjoint sgnt salt /\\ disjoint sgnt skey /\\\n disjoint salt msg /\\\n\n LS.rsapss_skey_pre (v modBits) (v eBits) (v dBits) (as_seq h skey) /\\\n LS.rsapss_sign_pre a (v modBits) (v saltLen) (as_seq h salt) (v msgLen) (as_seq h msg))\n (ensures fun h0 eq_m h1 -> modifies (loc sgnt) h0 h1 /\\\n (eq_m, as_seq h1 sgnt) == LS.rsapss_sign_ a (v modBits) (v eBits) (v dBits)\n (as_seq h0 skey) (v saltLen) (as_seq h0 salt) (v msgLen) (as_seq h0 msg))\n\n\ninline_for_extraction noextract\nval rsapss_sign_:\n #t:limb_t\n -> ke:BE.exp t\n -> a:Hash.hash_alg{S.hash_is_supported a}\n -> modBits:modBits_t t ->\n rsapss_sign_st1 t ke a modBits\n\nlet rsapss_sign_ #t ke a modBits eBits dBits skey saltLen salt msgLen msg sgnt =\n push_frame ();\n [@inline_let] let bits : size_pos = bits t in\n let nLen = blocks modBits (size bits) in\n let m = create nLen (uint #t 0) in\n rsapss_sign_msg_to_bn a modBits saltLen salt msgLen msg m;\n let eq_b = rsapss_sign_compute_sgnt ke modBits eBits dBits skey m sgnt in\n pop_frame ();\n eq_b\n\n\ninline_for_extraction noextract\nlet rsapss_sign_st (t:limb_t) (ke:BE.exp t) (a:Hash.hash_alg{S.hash_is_supported a}) (modBits:modBits_t t) =\n let len = blocks modBits (size (bits t)) in\n eBits:size_t\n -> dBits:size_t{LS.skey_len_pre t (v modBits) (v eBits) (v dBits)}\n -> skey:lbignum t (2ul *! len +! blocks eBits (size (bits t)) +! blocks dBits (size (bits t)))\n -> saltLen:size_t\n -> salt:lbuffer uint8 saltLen\n -> msgLen:size_t\n -> msg:lbuffer uint8 msgLen\n -> sgnt:lbuffer uint8 (blocks modBits 8ul) ->\n Stack bool\n (requires fun h -> len == ke.BE.bn.BN.len /\\\n live h salt /\\ live h msg /\\ live h sgnt /\\ live h skey /\\\n disjoint sgnt salt /\\ disjoint sgnt msg /\\ disjoint sgnt salt /\\ disjoint sgnt skey /\\\n disjoint salt msg /\\\n\n LS.rsapss_skey_pre (v modBits) (v eBits) (v dBits) (as_seq h skey))\n (ensures fun h0 b h1 -> modifies (loc sgnt) h0 h1 /\\\n (b, as_seq h1 sgnt) == LS.rsapss_sign a (v modBits) (v eBits) (v dBits)\n (as_seq h0 skey) (v saltLen) (as_seq h0 salt) (v msgLen) (as_seq h0 msg) (as_seq h0 sgnt))\n\n\ninline_for_extraction noextract\nval rsapss_sign:\n #t:limb_t\n -> ke:BE.exp t\n -> a:Hash.hash_alg{S.hash_is_supported a}\n -> modBits:modBits_t t ->\n rsapss_sign_st t ke a modBits\n\nlet rsapss_sign #t ke a modBits eBits dBits skey saltLen salt msgLen msg sgnt =\n let hLen = RM.hash_len a in\n Math.Lemmas.pow2_lt_compat 61 32;\n Math.Lemmas.pow2_lt_compat 125 32;\n //assert (max_size_t < Hash.max_input_length a);\n\n let b =\n saltLen <=. 0xfffffffful -! hLen -! 8ul &&\n saltLen +! hLen +! 2ul <=. blocks (modBits -! 1ul) 8ul in\n\n if b then\n rsapss_sign_ ke a modBits eBits dBits skey saltLen salt msgLen msg sgnt\n else\n false\n\n\n\ninline_for_extraction noextract\nval bn_lt_pow2:\n #t:limb_t\n -> modBits:size_t{1 < v modBits}\n -> m:lbignum t (blocks modBits (size (bits t))) ->\n Stack bool\n (requires fun h -> live h m)\n (ensures fun h0 r h1 -> h0 == h1 /\\\n r == LS.bn_lt_pow2 (v modBits) (as_seq h0 m))\n\nlet bn_lt_pow2 #t modBits m =\n if not ((modBits -! 1ul) %. 8ul =. 0ul) then true\n else begin\n let get_bit = BN.bn_get_ith_bit (blocks modBits (size (bits t))) m (modBits -! 1ul) in\n BB.unsafe_bool_of_limb0 get_bit end\n\n\ninline_for_extraction noextract\nlet rsapss_verify_bn_st (t:limb_t) (ke:BE.exp t) (modBits:modBits_t t) =\n let len = blocks modBits (size (bits t)) in\n eBits:size_t{LS.pkey_len_pre t (v modBits) (v eBits)}\n -> pkey:lbignum t (2ul *! len +! blocks eBits (size (bits t)))\n -> m_def:lbignum t len\n -> s:lbignum t len ->\n Stack bool\n (requires fun h -> len == ke.BE.bn.BN.len /\\\n live h pkey /\\ live h m_def /\\ live h s /\\\n disjoint m_def pkey /\\ disjoint m_def s /\\ disjoint s pkey /\\\n LS.rsapss_pkey_pre (v modBits) (v eBits) (as_seq h pkey))\n (ensures fun h0 r h1 -> modifies (loc m_def) h0 h1 /\\\n (r, as_seq h1 m_def) == LS.rsapss_verify_bn (v modBits) (v eBits) (as_seq h0 pkey) (as_seq h0 m_def) (as_seq h0 s))\n\n\ninline_for_extraction noextract\nval rsapss_verify_bn: #t:limb_t -> ke:BE.exp t -> modBits:modBits_t t -> rsapss_verify_bn_st t ke modBits\nlet rsapss_verify_bn #t ke modBits eBits pkey m_def s =\n [@inline_let] let bits = size (bits t) in\n let nLen = blocks modBits bits in\n let eLen = blocks eBits bits in\n\n let n = sub pkey 0ul nLen in\n let r2 = sub pkey nLen nLen in\n let e = sub pkey (nLen +! nLen) eLen in\n\n let mask = BN.bn_lt_mask nLen s n in\n let h = ST.get () in\n SB.bn_lt_mask_lemma (as_seq h s) (as_seq h n);\n\n let res =\n if BB.unsafe_bool_of_limb mask then begin\n Math.Lemmas.pow2_le_compat (v bits * v nLen) (v modBits);\n SM.bn_precomp_r2_mod_n_lemma (v modBits - 1) (as_seq h n);\n\n let h0 = ST.get () in\n BE.mk_bn_mod_exp_precompr2 nLen ke.BE.exp_vt_precomp n r2 s eBits e m_def;\n let h1 = ST.get () in\n SD.bn_eval_inj (v nLen) (as_seq h1 m_def)\n (SE.bn_mod_exp_vartime_precompr2 (v nLen) (as_seq h0 n) (as_seq h0 r2)\n (as_seq h1 s) (v eBits) (as_seq h0 e));\n\n if bn_lt_pow2 modBits m_def then true\n else false end\n else false in\n res\n\n\ninline_for_extraction noextract\nlet rsapss_verify_bn_to_msg_st (t:limb_t) (a:Hash.hash_alg{S.hash_is_supported a}) (modBits:modBits_t t) =\n saltLen:size_t\n -> msgLen:size_t\n -> msg:lbuffer uint8 msgLen\n -> m:lbignum t (blocks modBits (size (bits t))) ->\n Stack bool\n (requires fun h ->\n live h msg /\\ live h m /\\ disjoint m msg /\\\n\n LS.rsapss_verify_pre a (v saltLen) (v msgLen) (as_seq h msg))\n (ensures fun h0 r h1 -> modifies0 h0 h1 /\\\n r == LS.rsapss_verify_bn_to_msg a (v modBits) (v saltLen) (v msgLen) (as_seq h0 msg) (as_seq h0 m))\n\n\ninline_for_extraction noextract\nval rsapss_verify_bn_to_msg:\n #t:limb_t\n -> a:Hash.hash_alg{S.hash_is_supported a}\n -> modBits:modBits_t t ->\n rsapss_verify_bn_to_msg_st t a modBits\n\nlet rsapss_verify_bn_to_msg #t a modBits saltLen msgLen msg m =\n push_frame ();\n [@inline_let] let bits : size_pos = bits t in\n [@inline_let] let numb : size_pos = numbytes t in\n let nLen = blocks modBits (size bits) in\n\n let emBits = modBits -! 1ul in\n let emLen = blocks emBits 8ul in\n [@inline_let] let mLen = blocks emLen (size numb) in\n let em = create emLen (u8 0) in\n\n LS.blocks_bits_lemma t (v emBits);\n LS.blocks_numb_lemma t (v emBits);\n assert (SD.blocks (v emBits) bits == v mLen);\n assert (numb * v mLen <= max_size_t);\n assert (v mLen <= v nLen);\n\n let m1 = sub m 0ul mLen in\n BN.bn_to_bytes_be emLen m1 em;\n let res = RP.pss_verify a saltLen msgLen msg emBits em in\n pop_frame ();\n res\n\n\ninline_for_extraction noextract\nlet rsapss_verify_compute_msg_st (t:limb_t) (ke:BE.exp t) (modBits:modBits_t t) =\n let len = blocks modBits (size (bits t)) in\n eBits:size_t{LS.pkey_len_pre t (v modBits) (v eBits)}\n -> pkey:lbignum t (2ul *! len +! blocks eBits (size (bits t)))\n -> sgnt:lbuffer uint8 (blocks modBits 8ul)\n -> m:lbignum t len ->\n Stack bool\n (requires fun h -> len == ke.BE.bn.BN.len /\\\n live h sgnt /\\ live h pkey /\\ live h m /\\\n disjoint m sgnt /\\ disjoint m pkey /\\\n as_seq h m == LSeq.create (v len) (uint #t 0) /\\\n LS.rsapss_pkey_pre (v modBits) (v eBits) (as_seq h pkey))\n (ensures fun h0 r h1 -> modifies (loc m) h0 h1 /\\\n (r, as_seq h1 m) == LS.rsapss_verify_compute_msg (v modBits) (v eBits) (as_seq h0 pkey) (as_seq h0 sgnt))\n\n\ninline_for_extraction noextract\nval rsapss_verify_compute_msg:\n #t:limb_t\n -> ke:BE.exp t\n -> modBits:modBits_t t ->\n rsapss_verify_compute_msg_st t ke modBits\n\nlet rsapss_verify_compute_msg #t ke modBits eBits pkey sgnt m =\n push_frame ();\n [@inline_let] let bits : size_pos = bits t in\n [@inline_let] let numb : size_pos = numbytes t in\n let nLen = blocks modBits (size bits) in\n let k = blocks modBits 8ul in\n\n let s = create nLen (uint #t 0) in\n LS.blocks_bits_lemma t (v modBits);\n LS.blocks_numb_lemma t (v modBits);\n assert (SD.blocks (v k) numb == v nLen);\n assert (numb * v nLen <= max_size_t);\n BN.bn_from_bytes_be k sgnt s;\n\n let b = rsapss_verify_bn #t ke modBits eBits pkey m s in\n pop_frame ();\n b\n\n\ninline_for_extraction noextract\nlet rsapss_verify_st1 (t:limb_t) (ke:BE.exp t) (a:Hash.hash_alg{S.hash_is_supported a}) (modBits:modBits_t t) =\n let len = blocks modBits (size (bits t)) in\n eBits:size_t{LS.pkey_len_pre t (v modBits) (v eBits)}\n -> pkey:lbignum t (2ul *! len +! blocks eBits (size (bits t)))\n -> saltLen:size_t\n -> sgnt:lbuffer uint8 (blocks modBits 8ul)\n -> msgLen:size_t\n -> msg:lbuffer uint8 msgLen ->\n Stack bool\n (requires fun h -> len == ke.BE.bn.BN.len /\\\n live h msg /\\ live h sgnt /\\ live h pkey /\\\n disjoint msg sgnt /\\ disjoint msg pkey /\\\n\n LS.rsapss_pkey_pre (v modBits) (v eBits) (as_seq h pkey) /\\\n LS.rsapss_verify_pre a (v saltLen) (v msgLen) (as_seq h msg))\n (ensures fun h0 r h1 -> modifies0 h0 h1 /\\\n r == LS.rsapss_verify_ a (v modBits) (v eBits) (as_seq h0 pkey)\n (v saltLen) (as_seq h0 sgnt) (v msgLen) (as_seq h0 msg))\n\n\ninline_for_extraction noextract\nval rsapss_verify_:\n #t:limb_t\n -> ke:BE.exp t\n -> a:Hash.hash_alg{S.hash_is_supported a}\n -> modBits:modBits_t t ->\n rsapss_verify_st1 t ke a modBits\n\nPartial implementation:\nlet rsapss_verify_ #t ke a modBits eBits pkey saltLen sgnt msgLen msg =\n \n"
},
{
"role": "assistant",
"content": "push_frame ();\n[@@ inline_let ]let bits:size_pos = bits t in\nlet nLen = blocks modBits (size bits) in\nlet m = create nLen (uint #t 0) in\nlet b = rsapss_verify_compute_msg ke modBits eBits pkey sgnt m in\nlet res = if b then rsapss_verify_bn_to_msg a modBits saltLen msgLen msg m else false in\npop_frame ();\nres"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: imm: Vale.Def.Types_s.nat8 -> src: Vale.X64.Machine_s.quad32\n -> FStar.Pervasives.Native.option Vale.X64.Machine_s.quad32\nEffect: Prims.Tot\n\nContext:\nmodule Vale.X64.Instructions_s\nopen FStar.Mul\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Two_s\nopen Vale.Def.Words.Four_s\nopen Vale.Def.Types_s\nopen Vale.X64.Machine_s\nopen Vale.X64.Instruction_s\nopen Vale.X64.CPU_Features_s\nopen Vale.X64.CryptoInstructions_s\nopen FStar.Seq.Base\n\nlet eval_Mov64 (src:nat64) : option nat64 = Some src\nval ins_Mov64 : instr_dep [out op64] [op64] PreserveFlags eval_Mov64\n\nlet eval_MovBe64 (src:nat64) : option nat64 = if movbe_enabled then Some (reverse_bytes_nat64 src) else None\nval ins_MovBe64 : instr_dep [out op64] [op64] PreserveFlags eval_MovBe64\n\nlet eval_Bswap64 (dst:nat64) : option nat64 = Some (reverse_bytes_nat64 dst)\nval ins_Bswap64 : instr_dep [inOut op64] [] PreserveFlags eval_Bswap64\n\nlet eval_Cmovc64 (dst src:nat64) (carry:bool) : option nat64 = Some (if carry then src else dst)\nval ins_Cmovc64 : instr_dep [inOut op64] [op64; opFlagsCf] PreserveFlags eval_Cmovc64\n\nlet eval_Add64 (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src in Some (sum >= pow2_64, sum % pow2_64)\nval ins_Add64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Add64\n\nlet eval_AddLea64 (src1 src2:nat64) : option nat64 = Some ((src1 + src2) % pow2_64)\nval ins_AddLea64 :instr_dep [out op64] [op64; op64] PreserveFlags eval_AddLea64\n\nlet eval_AddCarry64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_carry then 1 else 0) in\n Some (sum >= pow2_64, sum % pow2_64)\nval ins_AddCarry64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_AddCarry64\n\nlet eval_Adcx64_Adox64 (old_flag:bool) (dst src:nat64) : option (bool & nat64) =\n let sum = dst + src + (if old_flag then 1 else 0) in\n if adx_enabled then Some (sum >= pow2_64, sum % pow2_64) else None\nval ins_Adcx64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\nval ins_Adox64 : instr_dep [inOut opFlagsOf; inOut op64] [op64] PreserveFlags eval_Adcx64_Adox64\n\nlet eval_Sub64 (dst src:nat64) : option (bool & nat64) =\n let diff = dst - src in Some (diff < 0, diff % pow2_64)\nval ins_Sub64 : instr_dep [out opFlagsCf; inOut op64] [op64] HavocFlags eval_Sub64\n\nlet eval_Sbb64 (old_carry:bool) (dst src:nat64) : option (bool & nat64) =\n let diff = dst - (src + (if old_carry then 1 else 0)) in\n Some (diff < 0, diff % pow2_64) // We specify cf, but underspecify everything else (via HavocFlags)\nval ins_Sbb64 : instr_dep [inOut opFlagsCf; inOut op64] [op64] HavocFlags eval_Sbb64\n\nlet eval_Mul64 (rax src:nat64) : option (nat64 & nat64) =\n Some (FStar.UInt.mul_div #64 rax src, FStar.UInt.mul_mod #64 rax src)\nval ins_Mul64 : instr_dep [out (one64Reg rRdx); inOut (one64Reg rRax)] [op64] HavocFlags eval_Mul64\n\nlet eval_Mulx64 (rdx src:nat64) : option (nat64 & nat64) =\n let hi = FStar.UInt.mul_div #64 rdx src in\n let lo = FStar.UInt.mul_mod #64 rdx src in\n if bmi2_enabled then Some (hi, lo) else None\nval ins_Mulx64 : instr_dep [out op64; out op64] [one64Reg rRdx; op64] PreserveFlags eval_Mulx64\n\nlet eval_IMul64 (dst src:nat64) : option nat64 =\n Some (FStar.UInt.mul_mod #64 dst src)\nval ins_IMul64 : instr_dep [inOut op64] [op64] HavocFlags eval_IMul64\n\nlet eval_And64 (dst src:nat64) : option nat64 = Some (iand dst src)\nval ins_And64 : instr_dep [inOut op64] [op64] HavocFlags eval_And64\n\nlet eval_Xor64 (dst src:nat64) : option (nat64 & (bool & bool)) =\n Some (Vale.Def.Types_s.ixor dst src, (false, false))\nval ins_Xor64 : instr_dep [inOut op64; out opFlagsCf; out opFlagsOf] [op64] HavocFlags eval_Xor64\n\nlet eval_Shr64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishr dst amt) else None\nval ins_Shr64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shr64\n\nlet eval_Shl64 (dst amt:nat64) : option nat64 =\n if amt < 64 then Some (Vale.Def.Types_s.ishl dst amt) else None\nval ins_Shl64 : instr_dep [inOut op64] [op64] HavocFlags eval_Shl64\n\nlet eval_Cpuid (rax rcx:nat64) : option (nat64 & (nat64 & (nat64 & nat64))) =\n Some (cpuid rRax rax rcx, (cpuid rRbx rax rcx, (cpuid rRcx rax rcx, cpuid rRdx rax rcx)))\nval ins_Cpuid :\n instr_dep\n [inOut (one64Reg rRax); out (one64Reg rRbx); inOut (one64Reg rRcx); out (one64Reg rRdx)]\n [] PreserveFlags eval_Cpuid\n\n// The XGETBV instruction requires that OSXSAVE (in CPUID) is enabled.\n// We underspecify XGETBV here to only support fetching XCR0, which\n// is supported on any processor supporting the XGETBV instruction\nlet eval_Xgetbv (rcx:nat64) : option (nat64 & nat64) =\n if osxsave_enabled && rcx = 0 then Some (xgetbv rRax rcx, xgetbv rRdx rcx) else None\nval ins_Xgetbv :\n instr_dep\n [out (one64Reg rRax); out (one64Reg rRdx)]\n [one64Reg rRcx] PreserveFlags eval_Xgetbv\n\nlet check_avx (#a:Type0) (x:option a) : option a =\n if avx_enabled then x else None\n\nlet check_sse2 (#a:Type0) (x:option a) : option a =\n if sse2_enabled then x else None\n\nlet check_ssse3 (#a:Type0) (x:option a) : option a =\n if ssse3_enabled then x else None\n\nlet check_sse4_1 (#a:Type0) (x:option a) : option a =\n if sse4_1_enabled then x else None\n\nlet eval_Movdqu (src:quad32) : option quad32 = check_sse2 (Some src)\nval ins_Movdqu : instr_dep [out opXmm] [opXmm] PreserveFlags eval_Movdqu\n\nlet eval_Pxor (dst src:quad32) : option quad32 = check_sse2 (Some (quad32_xor dst src))\nval ins_Pxor : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pxor\n\nlet eval_VPxor (src1 src2:quad32) : option quad32 = check_avx (Some (quad32_xor src1 src2))\nval ins_VPxor : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPxor\n\nlet eval_Pand (dst src:quad32) : option quad32 = check_sse2 (Some (four_map2 (fun di si -> iand di si) dst src))\nval ins_Pand : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pand\n\nlet eval_Paddd_raw (src1 src2:quad32) : option quad32 =\n Some (Mkfour\n ((src1.lo0 + src2.lo0) % pow2_32)\n ((src1.lo1 + src2.lo1) % pow2_32)\n ((src1.hi2 + src2.hi2) % pow2_32)\n ((src1.hi3 + src2.hi3) % pow2_32))\n\nlet eval_Paddd (src1 src2:quad32) : option quad32 = check_sse2 (eval_Paddd_raw src1 src2)\nval ins_Paddd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Paddd\n\nlet eval_VPaddd (src1 src2:quad32) : option quad32 = check_avx (eval_Paddd_raw src1 src2)\nval ins_VPaddd : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPaddd\n\nlet eval_Pslld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishl i amt) dst) else None)\nval ins_Pslld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Pslld amt)\n\nlet eval_Psrld (amt:int) (dst:quad32) : option quad32 =\n check_sse2 (if 0 <= amt && amt < 32 then Some (four_map (fun i -> ishr i amt) dst) else None)\nval ins_Psrld (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrld amt)\n\nlet eval_Psrldq (amt:int) (dst:quad32) : option quad32 = check_sse2 (\n if 0 <= amt && amt < 16 then\n let src_bytes = le_quad32_to_bytes dst in\n let zero_pad = Seq.create amt 0 in\n let remaining_bytes = slice src_bytes amt (length src_bytes) in\n Some (le_bytes_to_quad32 (append zero_pad remaining_bytes))\n else None)\nval ins_Psrldq (amt:int) : instr_dep [inOut opXmm] [] PreserveFlags (eval_Psrldq amt)\n\nlet eval_Palignr_raw (amount:nat8) (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if amount = 4 then\n Some (Mkfour src2.lo1 src2.hi2 src2.hi3 src1.lo0)\n else if amount = 8 then\n Some (Mkfour src2.hi2 src2.hi3 src1.lo0 src1.lo1)\n else None\n\nlet eval_Palignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_ssse3 (eval_Palignr_raw amount src1 src2)\nval ins_Palignr (amount:nat8) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Palignr amount)\n\nlet eval_VPalignr (amount:nat8) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Palignr_raw amount src1 src2)\nval ins_VPalignr (amount:nat8) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPalignr amount)\n\nlet eval_Shufpd_raw (permutation:int) (src1 src2:quad32) : option quad32 =\n if 0 <= permutation && permutation < 4 then\n Some (Mkfour\n (if permutation % 2 = 0 then src1.lo0 else src1.hi2)\n (if permutation % 2 = 0 then src1.lo1 else src1.hi3)\n (if (permutation / 2) % 2 = 0 then src2.lo0 else src2.hi2)\n (if (permutation / 2) % 2 = 0 then src2.lo1 else src2.hi3))\n else None\n\nlet eval_Shufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_sse2 (eval_Shufpd_raw permutation src1 src2)\nval ins_Shufpd (permutation:int) :\n instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Shufpd permutation)\n\nlet eval_VShufpd (permutation:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Shufpd_raw permutation src1 src2)\nval ins_VShufpd (permutation:int) :\n instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VShufpd permutation)\n\nlet is_full_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x04050607 &&\n q.hi3 = 0x00010203\n\nlet is_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x00010203 &&\n q.lo1 = 0x04050607 &&\n q.hi2 = 0x08090A0B &&\n q.hi3 = 0x0C0D0E0F\n\nlet is_high_dup_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x0C0D0E0F &&\n q.lo1 = 0x08090A0B &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet is_lower_upper_byte_reversal_mask (q:quad32) : bool =\n q.lo0 = 0x04050607 &&\n q.lo1 = 0x00010203 &&\n q.hi2 = 0x0C0D0E0F &&\n q.hi3 = 0x08090A0B\n\nlet eval_Pshufb_raw (src1 src2:quad32) : option quad32 =\n // We only spec a restricted version sufficient for a handful of standard patterns\n if is_full_byte_reversal_mask src2 then\n Some (reverse_bytes_quad32 src1)\n else if is_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3))\n else if is_high_dup_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else if is_lower_upper_byte_reversal_mask src2 then\n Some (Mkfour\n (reverse_bytes_nat32 src1.lo1)\n (reverse_bytes_nat32 src1.lo0)\n (reverse_bytes_nat32 src1.hi3)\n (reverse_bytes_nat32 src1.hi2))\n else None\n\nlet eval_Pshufb (src1 src2:quad32) : option quad32 = check_ssse3 (eval_Pshufb_raw src1 src2)\nval ins_Pshufb : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pshufb\n\nlet eval_VPshufb (src1 src2:quad32) : option quad32 = check_avx (eval_Pshufb_raw src1 src2)\nval ins_VPshufb : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VPshufb\n\nlet eval_Pshufd (permutation:nat8) (src:quad32) : option quad32 = check_sse2 (\n let bits:bits_of_byte = byte_to_twobits permutation in\n Some (Mkfour\n (select_word src bits.lo0)\n (select_word src bits.lo1)\n (select_word src bits.hi2)\n (select_word src bits.hi3)))\nval ins_Pshufd (permutation:nat8) :\n instr_dep [out opXmm] [opXmm] PreserveFlags (eval_Pshufd permutation)\n\nlet eval_Pcmpeqd (dst src:quad32) : option quad32 = check_sse2 (\n let eq_result (b:bool):nat32 = if b then 0xFFFFFFFF else 0 in\n Some (Mkfour\n (eq_result (src.lo0 = dst.lo0))\n (eq_result (src.lo1 = dst.lo1))\n (eq_result (src.hi2 = dst.hi2))\n (eq_result (src.hi3 = dst.hi3))))\nval ins_Pcmpeqd : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_Pcmpeqd\n\nlet eval_Pextrq (index:nat8) (src:quad32) : option nat64 = check_sse4_1 (\n let src_two = four_to_two_two src in\n Some (two_to_nat 32 (two_select src_two (index % 2))))\nval ins_Pextrq (index:nat8) : instr_dep [out op64] [opXmm] PreserveFlags (eval_Pextrq index)\n\nlet eval_Pinsrd (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat32 dst (src % pow2_32) (index % 4)))\nval ins_Pinsrd (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrd index)\n\nlet eval_Pinsrq (index:nat8) (dst:quad32) (src:nat64) : option quad32 =\n check_sse4_1 (Some (insert_nat64_def dst src (index % 2)))\nval ins_Pinsrq (index:nat8) : instr_dep [inOut opXmm] [op64] PreserveFlags (eval_Pinsrq index)\n\nlet eval_Pslldq_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the two very special cases we need\n if count = 4 then Some (Mkfour 0 src.lo0 src.lo1 src.hi2)\n else if count = 8 then Some (Mkfour 0 0 src.lo0 src.lo1)\n else None\nlet eval_VPslldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Pslldq_raw count src)\nval ins_VPslldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPslldq count)\n\nlet eval_Psrldq_8_raw (count:nat8) (src:quad32) : option quad32 =\n // We only spec the one very special case we need\n if count = 8 then Some (Mkfour src.hi2 src.hi3 0 0)\n else None\nlet eval_VPsrldq (count:nat8) (src:quad32) : option quad32 = check_avx (eval_Psrldq_8_raw count src)\nval ins_VPsrldq (count:nat8) : instr_dep [out opXmm] [opXmm] PreserveFlags (eval_VPsrldq count)\n\nlet eval_Pclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n let Mkfour a0 a1 a2 a3 = src1 in\n let Mkfour b0 b1 b2 b3 = src2 in\n let f x0 x1 y0 y1 =\n let x = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo x0 x1) in\n let y = Vale.Math.Poly2.Bits_s.of_double32 (Mktwo y0 y1) in\n Vale.Math.Poly2.Bits_s.to_quad32 (Vale.Math.Poly2_s.mul x y)\n in\n if pclmulqdq_enabled then\n match imm with\n | 0 -> Some (f a0 a1 b0 b1)\n | 1 -> Some (f a2 a3 b0 b1)\n | 16 -> Some (f a0 a1 b2 b3)\n | 17 -> Some (f a2 a3 b2 b3)\n | _ -> None\n else None\nval ins_Pclmulqdq (imm:int) : instr_dep [inOut opXmm] [opXmm] PreserveFlags (eval_Pclmulqdq imm)\n\nlet eval_VPclmulqdq (imm:int) (src1 src2:quad32) : option quad32 =\n check_avx (eval_Pclmulqdq imm src1 src2)\nval ins_VPclmulqdq (imm:int) : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags (eval_VPclmulqdq imm)\n\nlet eval_AESNI_enc (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.mix_columns_LE (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1))) src2)\n else None\nval ins_AESNI_enc : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc\n\nlet eval_VAESNI_enc (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc src1 src2)\nval ins_VAESNI_enc : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc\n\nlet eval_AESNI_enc_last (src1 src2:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.sub_bytes (Vale.AES.AES_s.shift_rows_LE src1)) src2)\n else None\nval ins_AESNI_enc_last : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_enc_last\n\nlet eval_VAESNI_enc_last (src1 src2:quad32) : option quad32 = check_avx (eval_AESNI_enc_last src1 src2)\nval ins_VAESNI_enc_last : instr_dep [out opXmm] [opXmm; opXmm] PreserveFlags eval_VAESNI_enc_last\n\nlet eval_AESNI_dec (dst src:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.inv_mix_columns_LE (Vale.AES.AES_s.inv_sub_bytes (Vale.AES.AES_s.inv_shift_rows_LE dst))) src)\n else None\nval ins_AESNI_dec : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_dec\n\nlet eval_AESNI_dec_last (dst src:quad32) : option quad32 =\n if aesni_enabled then\n Some (quad32_xor (Vale.AES.AES_s.inv_sub_bytes (Vale.AES.AES_s.inv_shift_rows_LE dst)) src)\n else None\nval ins_AESNI_dec_last : instr_dep [inOut opXmm] [opXmm] PreserveFlags eval_AESNI_dec_last\n\nlet eval_AESNI_imc (src:quad32) : option quad32 =\n if aesni_enabled then Some (Vale.AES.AES_s.inv_mix_columns_LE src) else None\nval ins_AESNI_imc : instr_dep [out opXmm] [opXmm] PreserveFlags eval_AESNI_imc\n\nPartial implementation:\nlet eval_AESNI_keygen_assist (imm: nat8) (src: quad32) : option quad32 =\n \n"
},
{
"role": "assistant",
"content": "if aesni_enabled\nthen\n Some\n (Mkfour (Vale.AES.AES_s.sub_word src.lo1)\n (ixor (Vale.AES.AES_s.rot_word_LE (Vale.AES.AES_s.sub_word src.lo1)) imm)\n (Vale.AES.AES_s.sub_word src.hi3)\n (ixor (Vale.AES.AES_s.rot_word_LE (Vale.AES.AES_s.sub_word src.hi3)) imm))\nelse None"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: a: Type -> Type\nEffect: Prims.Tot\n\nContext:\n(*\n Copyright 2020 Microsoft Research\n\n Licensed under the Apache License, Version 2.0 (the \"License\");\n you may not use this file except in compliance with the License.\n You may obtain a copy of the License at\n\n http://www.apache.org/licenses/LICENSE-2.0\n\n Unless required by applicable law or agreed to in writing, software\n distributed under the License is distributed on an \"AS IS\" BASIS,\n WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n See the License for the specific language governing permissions and\n limitations under the License.\n*)\n\nmodule Steel.GhostMonotonicHigherReference\n\nopen FStar.PCM\nopen FStar.Ghost\nopen Steel.FractionalPermission\n\nopen Steel.Memory\nopen Steel.Effect.Atomic\nopen Steel.Effect\n\nmodule Preorder = FStar.Preorder\n\n/// A library for Steel references that are monotonic with respect to a user-specified preorder.\n/// This library builds on top of Steel.HigherReference, and is specialized to values at universe 1.\n\n/// An abstract datatype for monotonic references\n[@@erasable]\nval ref (a:Type u#1) (p:Preorder.preorder a)\n : Type u#0\n\n/// The standard points to separation logic predicate\nval pts_to_sl (#a:Type) (#p:Preorder.preorder a) (r:ref a p) (f:perm) (v:a)\n : slprop u#1\n\n/// Lifting the standard points to predicate to vprop, with a non-informative selector\n[@@ __steel_reduce__]\nunfold let pts_to (#a:Type) (#p:Preorder.preorder a) (r:ref a p) (f:perm) (v:a) : vprop =\n to_vprop (pts_to_sl r f v)\n\n/// Allocates a reference with value [x]. We have full permission on the newly\n/// allocated reference.\nval alloc (#opened: _) (#a:Type) (p:Preorder.preorder a) (v:a)\n : SteelGhostT (ref a p) opened emp (fun r -> pts_to r full_perm v)\n\n/// Writes value [x] in the reference [r], as long as we have full ownership of [r]\nval write (#opened: _) (#a:Type) (#p:Preorder.preorder a) (#v:a)\n (r:ref a p) (x:a)\n : SteelGhost unit opened (pts_to r full_perm v)\n (fun v -> pts_to r full_perm x)\n (requires fun _ -> p v x /\\ True)\n (ensures fun _ _ _ -> True)\n\nPartial implementation:\nlet property (a: Type) =\n \n"
},
{
"role": "assistant",
"content": "a -> prop"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: p: FStar.Preorder.preorder a -> Type\nEffect: Prims.Tot\n\nContext:\n(*\n Copyright 2020 Microsoft Research\n\n Licensed under the Apache License, Version 2.0 (the \"License\");\n you may not use this file except in compliance with the License.\n You may obtain a copy of the License at\n\n http://www.apache.org/licenses/LICENSE-2.0\n\n Unless required by applicable law or agreed to in writing, software\n distributed under the License is distributed on an \"AS IS\" BASIS,\n WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n See the License for the specific language governing permissions and\n limitations under the License.\n*)\n\nmodule Steel.GhostMonotonicHigherReference\n\nopen FStar.PCM\nopen FStar.Ghost\nopen Steel.FractionalPermission\n\nopen Steel.Memory\nopen Steel.Effect.Atomic\nopen Steel.Effect\n\nmodule Preorder = FStar.Preorder\n\n/// A library for Steel references that are monotonic with respect to a user-specified preorder.\n/// This library builds on top of Steel.HigherReference, and is specialized to values at universe 1.\n\n/// An abstract datatype for monotonic references\n[@@erasable]\nval ref (a:Type u#1) (p:Preorder.preorder a)\n : Type u#0\n\n/// The standard points to separation logic predicate\nval pts_to_sl (#a:Type) (#p:Preorder.preorder a) (r:ref a p) (f:perm) (v:a)\n : slprop u#1\n\n/// Lifting the standard points to predicate to vprop, with a non-informative selector\n[@@ __steel_reduce__]\nunfold let pts_to (#a:Type) (#p:Preorder.preorder a) (r:ref a p) (f:perm) (v:a) : vprop =\n to_vprop (pts_to_sl r f v)\n\n/// Allocates a reference with value [x]. We have full permission on the newly\n/// allocated reference.\nval alloc (#opened: _) (#a:Type) (p:Preorder.preorder a) (v:a)\n : SteelGhostT (ref a p) opened emp (fun r -> pts_to r full_perm v)\n\n/// Writes value [x] in the reference [r], as long as we have full ownership of [r]\nval write (#opened: _) (#a:Type) (#p:Preorder.preorder a) (#v:a)\n (r:ref a p) (x:a)\n : SteelGhost unit opened (pts_to r full_perm v)\n (fun v -> pts_to r full_perm x)\n (requires fun _ -> p v x /\\ True)\n (ensures fun _ _ _ -> True)\n\n/// A wrapper around a predicate that depends on a value of type [a]\nlet property (a:Type)\n = a -> prop\n\n/// A wrapper around a property [fact] that has been witnessed to be true and stable\n/// with respect to preorder [p]\nval witnessed (#a:Type u#1) (#p:Preorder.preorder a) (r:ref a p) (fact:property a)\n : Type0\n\n/// The type of properties depending on values of type [a], and that\n\nPartial implementation:\nlet stable_property (#a: Type) (p: Preorder.preorder a) =\n \n"
},
{
"role": "assistant",
"content": "fact: property a {Preorder.stable fact p}"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: r: Steel.GhostMonotonicHigherReference.ref a p -> f: Steel.FractionalPermission.perm -> v: a\n -> Steel.Effect.Common.vprop\nEffect: Prims.Tot\n\nContext:\n(*\n Copyright 2020 Microsoft Research\n\n Licensed under the Apache License, Version 2.0 (the \"License\");\n you may not use this file except in compliance with the License.\n You may obtain a copy of the License at\n\n http://www.apache.org/licenses/LICENSE-2.0\n\n Unless required by applicable law or agreed to in writing, software\n distributed under the License is distributed on an \"AS IS\" BASIS,\n WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n See the License for the specific language governing permissions and\n limitations under the License.\n*)\n\nmodule Steel.GhostMonotonicHigherReference\n\nopen FStar.PCM\nopen FStar.Ghost\nopen Steel.FractionalPermission\n\nopen Steel.Memory\nopen Steel.Effect.Atomic\nopen Steel.Effect\n\nmodule Preorder = FStar.Preorder\n\n/// A library for Steel references that are monotonic with respect to a user-specified preorder.\n/// This library builds on top of Steel.HigherReference, and is specialized to values at universe 1.\n\n/// An abstract datatype for monotonic references\n[@@erasable]\nval ref (a:Type u#1) (p:Preorder.preorder a)\n : Type u#0\n\n/// The standard points to separation logic predicate\nval pts_to_sl (#a:Type) (#p:Preorder.preorder a) (r:ref a p) (f:perm) (v:a)\n : slprop u#1\n\n/// Lifting the standard points to predicate to vprop, with a non-informative selector\n\nPartial implementation:\nlet pts_to (#a: Type) (#p: Preorder.preorder a) (r: ref a p) (f: perm) (v: a) : vprop =\n \n"
},
{
"role": "assistant",
"content": "to_vprop (pts_to_sl r f v)"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: h: FStar.Monotonic.HyperStack.mem -> c: LowStar.Buffer.pointer (LowStar.Lib.LinkedList.cell a)\n -> Prims.GTot a\nEffect: Prims.GTot\n\nContext:\nmodule LowStar.Lib.LinkedList\n\n/// This module, oftentimes referred to as LL1 in other parts of the\n/// documentation, provides a low-level view over a linked list. Unless you\n/// intend on modifying the structure of the linked list (e.g. removing cells)\n/// or iterating over each cell, your are probably better off using LL2.\n///\n/// This module is intentionally not relying on a tight abstraction, to i) permit\n/// direct-style iteration by clients and ii) allow stating some SMTPat lemmas\n/// only on LL1 rather than replicating them in LL2 (going from LL2 to LL1 can\n/// happen automatically at the SMT level since LL2 does not have an fsti either).\n///\n/// Some open questions relative to the design of this module:\n/// - should ``v`` be defined directly in terms of ``cells`` rather than have a\n/// recursive definition along with an equivalence lemma?\n/// - should ``footprint`` be defined directly in terms of ``cells`` rather than\n/// have a recursive definition along with an equivalence lemma?\n/// - is it worth the trouble defining a footprint? my intuition is that it's\n/// fine and allows some clients to do very precise reasoning -- other modules\n/// (e.g. AssocList), when establishing abstraction boundaries, can simplify\n/// things and get rid of the precise footprint reasoning in favor of a static\n/// region-based footprint\n///\n/// Some remaining work items:\n/// - switch to C.Loops.while\n\nopen LowStar.BufferOps\n\nmodule B = LowStar.Buffer\nmodule HS = FStar.HyperStack\nmodule G = FStar.Ghost\nmodule L = FStar.List.Tot\nmodule U32 = FStar.UInt32\nmodule ST = FStar.HyperStack.ST\n\nopen FStar.HyperStack.ST\n\n/// Breaking from my own recommendation here... using fuel 1 / ifuel 1 since\n/// everything does recursive list reasoning.\n#set-options \"--__no_positivity --fuel 1 --ifuel 1\"\n\n/// Definition of the Low* type\n/// ---------------------------\nnoeq\ntype t (a: Type0) =\n B.pointer_or_null (cell a)\n\nand cell (a: Type0) = {\n next: t a;\n data: a;\n}\n\n/// Unlike the canonical style that associates a ``v`` function to any stateful\n/// representation, we use a relation here. Writing a function is just\n/// impossible, since we can't prove termination of ``v`` owing to the fact that\n/// there may be an infinite number of cells in the heap.\n///\n/// LL2 is the module that switches to the function style.\n///\n/// Note: no need to use erased here for ``l`` since the function is already in ``GTot``.\nlet rec well_formed #a (h: HS.mem) (c: t a) (l: list a): GTot Type0 (decreases l) =\n B.live h c /\\ (\n match l with\n | [] ->\n B.g_is_null c\n | a :: q ->\n B.length c == 1 /\\ (\n let { data=data; next=next } = B.get h c 0 in\n a == data /\\\n well_formed h next (G.hide q)\n ))\n\n/// Precise reasoning via the cells predicate\n/// -----------------------------------------\n///\n/// The modifies clauses are given either on the footprint of the list, or on a\n/// region that includes the footprint of the list (higher up the stack, e.g.\n/// LL2). This is good, but does not rule out \"stupid\" implementations that\n/// might re-allocate every single list cell when, say, doing a pop.\n///\n/// For advanced usages which take pointers directly to individual list cells,\n/// it's important to rule out these cases. We thus define the ``cells``\n/// predicate for that purpose.\n\nlet rec cells #a (h: HS.mem) (c: t a) (l: list a): Ghost (list (B.pointer (cell a)))\n (requires well_formed h c l)\n (ensures fun _ -> True)\n (decreases l)\n=\n if B.g_is_null c then\n []\n else\n c :: cells h (B.deref h c).next (List.Tot.tl l)\n\n/// A missing bit of automated reasoning\n/// ------------------------------------\n\n/// This should be absolutely trivial, yet, because there's no pattern on\n/// null_unique, we have to do the case analysis and the decomposition\n/// ourselves. This is quite tedious, as it usually takes a while (at least it\n/// did for me!) to realize that this property does not hold automatically.\n///\n/// I'm setting an SMTPat on this lemma because it's absolutely impossible to\n/// get any work done without it.\nlet same_cells_same_pointer #a (h0 h1: HS.mem) (ll0 ll1: t a) (l0 l1: list a): Lemma\n (requires (\n well_formed h0 ll0 l0 /\\\n well_formed h1 ll1 l1 /\\\n cells h0 ll0 l0 == cells h1 ll1 l1))\n (ensures (\n ll0 == ll1))\n [ SMTPat (cells h0 ll0 l0); SMTPat (cells h1 ll1 l1) ]\n=\n if B.g_is_null ll0 && B.g_is_null ll1 then begin\n B.null_unique ll0;\n B.null_unique ll1\n end else if not (B.g_is_null ll0) && not (B.g_is_null ll1) then begin\n ()\n end else\n false_elim ()\n\n/// Classic stateful reasoning lemmas\n/// ---------------------------------\n///\n/// This essentially follows ``stateful``, the type class of a mutable piece of\n/// state defined using the modifies-theory of Low*v2 and equipped with the\n/// right lemmas to enable usage by clients. See\n/// hacl-star/code/streaming/Hacl.Streaming.Interface.fst for more context.\n\nval footprint: (#a: Type) -> (h: HS.mem) -> (l: t a) -> (n: list a) -> Ghost B.loc\n (requires (well_formed h l n))\n (ensures (fun refs -> True))\n (decreases n)\n\nlet rec footprint #a h l n =\n if B.g_is_null l\n then B.loc_none\n else\n let {next = next} = B.get h l 0 in\n let refs = footprint h next (G.hide (L.tl n)) in\n B.loc_union (B.loc_addr_of_buffer l) refs\n\n/// Departing from LinkedList4 here. I prefer to bolt these into the invariant\n/// rather than requiring on some wizard lemmas from LowStar.Monotonic.Buffer to\n/// deduce these.\nval cells_pairwise_disjoint: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_pairwise_disjoint #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.loc_disjoint (B.loc_addr_of_buffer l) (footprint h next (L.tl n)) /\\\n cells_pairwise_disjoint h next (L.tl n)\n\nval cells_live_freeable: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_live_freeable #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.live h l /\\ B.freeable l /\\\n cells_live_freeable h next (L.tl n)\n\nval invariant: #a:Type -> h:HS.mem -> l: t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\nlet invariant #a h l n =\n cells_live_freeable h l n /\\ cells_pairwise_disjoint h l n\n\n/// Normally this would be automatic and writing a custom lemma for that stuff\n/// would be needed only when writing an fsti, however, as stated earlier,\n/// recursion makes everything more difficult and this absolutely needs to be in\n/// scope otherwise the modifies-clause theory cannot deduce disjointness of\n/// fresh allocations w.r.t. the footprint.\nlet rec invariant_loc_in_footprint (#a: Type) (h: HS.mem) (l: t a) (n: list a): Lemma\n (requires well_formed h l n)\n (ensures B.loc_in (footprint h l n) h)\n (decreases n)\n [ SMTPat (invariant h l n) ]\n=\n if B.g_is_null l then\n ()\n else\n let next = (B.deref h l).next in\n invariant_loc_in_footprint h next (L.tl n)\n\n/// Another absolutely essential lemma, for which interestingly the pattern\n/// \"invariant h1 l n\" does not work, but well_formed does... I guess that's\n/// alright.\nlet rec frame (#a: Type) (l: t a) (n: list a) (r: B.loc) (h0 h1: HS.mem): Lemma\n (requires (\n well_formed h0 l n /\\\n invariant h0 l n /\\\n B.loc_disjoint r (footprint h0 l n) /\\\n B.modifies r h0 h1\n ))\n (ensures (\n well_formed h1 l n /\\\n footprint h1 l n == footprint h0 l n /\\\n cells h1 l n == cells h0 l n /\\\n invariant h1 l n\n ))\n (decreases n)\n [ SMTPat (well_formed h1 l n); SMTPat (B.modifies r h0 h1) ]\n=\n if B.g_is_null l then\n ()\n else\n frame (B.deref h0 l).next (L.tl n) r h0 h1\n\n/// Lemmas for working with linked lists\n/// ------------------------------------\n///\n/// Since it's a recursive data structure, a lot of the automated reasoning from\n/// LowStar.Buffer doesn't work since it requires inductions. Some helpful\n/// properties are therefore proved here.\n\nlet rec length_functional #a (h: HS.mem) (c: t a) (l1 l2: list a):\n Lemma\n (requires (well_formed h c l1 /\\ well_formed h c l2))\n (ensures (l1 == l2))\n (decreases (G.reveal l1))\n [ SMTPat (well_formed h c l1); SMTPat (well_formed h c l2) ] =\n if B.g_is_null c\n then ()\n else\n let { next=next } = B.get h c 0 in\n length_functional h next (G.hide (L.tl l1)) (G.hide (L.tl l2))\n\n/// The footprint is based on loc_addr_of_buffer so that we can write a free\n/// operation that operates on the footprint. However, this invalidates most of\n/// the helper lemmas for disjointness that were present in the previous\n/// iteration of this module, named LinkedList4.fst. It's ok, we bake\n/// disjointness directly onto the invariant and it works just as well.\n\n/// Connection between ``cells`` and ``v``\n/// --------------------------------------\n\n/// Redefining some helpers.\n#push-options \"--ifuel 1\"\nlet rec gmap #a #b (f: a -> GTot b) (xs: list a): GTot (list b) =\n match xs with\n | [] -> []\n | x :: xs -> f x :: gmap f xs\n\n/// \"As we all know\", right folds are the easiest to work with because they\n/// follow the natural structure of recursion (left-folds are evil).\nlet rec gfold_right #a #b (f: b -> a -> GTot b) (xs: list a) (acc: b): Ghost b\n (requires True)\n (ensures fun _ -> True)\n (decreases xs)\n=\n match xs with\n | [] -> acc\n | x :: xs -> f (gfold_right f xs acc) x\n#pop-options\n\n/// Connecting ``v`` and ``cells``.\n///\n/// See ``tests/Wireguard.fst`` for why going through gmap is important and how\n/// to structure your definitions (should you need to reason about ``cells``) in\n\nPartial implementation:\nlet deref_data #a (h: HS.mem) (c: B.pointer (cell a)) : GTot a =\n \n"
},
{
"role": "assistant",
"content": "(B.deref h c).data"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n alg: Vale.AES.AES_common_s.algorithm ->\n offset: Prims.int ->\n auth_b: Vale.X64.Memory.buffer128 ->\n abytes_b: Vale.X64.Memory.buffer128 ->\n in128x6_b: Vale.X64.Memory.buffer128 ->\n out128x6_b: Vale.X64.Memory.buffer128 ->\n in128_b: Vale.X64.Memory.buffer128 ->\n out128_b: Vale.X64.Memory.buffer128 ->\n inout_b: Vale.X64.Memory.buffer128 ->\n iv_b: Vale.X64.Memory.buffer128 ->\n iv: Vale.AES.GCM_s.supported_iv_LE ->\n scratch_b: Vale.X64.Memory.buffer128 ->\n key: FStar.Seq.Base.seq Vale.X64.Memory.nat32 ->\n round_keys: FStar.Seq.Base.seq Vale.X64.Decls.quad32 ->\n keys_b: Vale.X64.Memory.buffer128 ->\n hkeys_b: Vale.X64.Memory.buffer128 ->\n expected_tag: FStar.Seq.Base.seq Vale.X64.Memory.nat8\n -> Vale.X64.QuickCode.va_quickCode Prims.unit\n (Vale.AES.X64.GCMdecryptOpt.va_code_Gcm_blocks_wrapped alg offset)\nEffect: Prims.Tot\n\nContext:\nmodule Vale.AES.X64.GCMdecryptOpt\nopen Vale.Def.Prop_s\nopen Vale.Def.Opaque_s\nopen FStar.Seq\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Seq_s\nopen Vale.Def.Types_s\nopen Vale.Arch.Types\nopen Vale.Arch.HeapImpl\nopen Vale.AES.AES_s\nopen Vale.AES.GCTR_s\nopen Vale.AES.GCTR\nopen Vale.AES.GCM\nopen Vale.AES.GHash_s\nopen Vale.AES.GHash\nopen Vale.AES.GCM_s\nopen Vale.AES.X64.AES\nopen Vale.AES.GF128_s\nopen Vale.AES.GF128\nopen Vale.Poly1305.Math\nopen Vale.AES.GCM_helpers\nopen Vale.AES.X64.GHash\nopen Vale.AES.X64.GCTR\nopen Vale.X64.Machine_s\nopen Vale.X64.Memory\nopen Vale.X64.Stack_i\nopen Vale.X64.State\nopen Vale.X64.Decls\nopen Vale.X64.InsBasic\nopen Vale.X64.InsMem\nopen Vale.X64.InsVector\nopen Vale.X64.InsStack\nopen Vale.X64.InsAes\nopen Vale.X64.QuickCode\nopen Vale.X64.QuickCodes\nopen Vale.AES.X64.GF128_Mul\nopen Vale.X64.Stack\nopen Vale.X64.CPU_Features_s\nopen Vale.Math.Poly2.Bits_s\nopen Vale.AES.X64.AESopt\nopen Vale.AES.X64.AESGCM\nopen Vale.AES.X64.AESopt2\nopen Vale.Lib.Meta\nopen Vale.AES.X64.GCMencryptOpt\nopen Vale.AES.OptPublic\nopen Vale.Lib.Basic\n#reset-options \"--z3rlimit 20 --max_ifuel 0\"\n//-- Gcm_extra_bytes\n\nval va_code_Gcm_extra_bytes : alg:algorithm -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_extra_bytes alg =\n (va_Block (va_CCons (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0)\n (va_op_reg_opr64_reg64 rRax) 0 Secret) (va_CCons (va_code_Mov128 (va_op_xmm_xmm 10)\n (va_op_xmm_xmm 0)) (va_CCons (va_code_Ghash_extra_bytes ()) (va_CCons (va_code_Mov128\n (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_CCons (va_code_Pshufb (va_op_xmm_xmm 0)\n (va_op_xmm_xmm 9)) (va_CCons (va_code_AESEncryptBlock alg) (va_CCons (va_code_Pxor\n (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (va_CCons (va_code_Store128_buffer\n (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax) (va_op_xmm_xmm 10) 0 Secret)\n (va_CNil ()))))))))))\n\nval va_codegen_success_Gcm_extra_bytes : alg:algorithm -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_extra_bytes alg =\n (va_pbool_and (va_codegen_success_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0)\n (va_op_reg_opr64_reg64 rRax) 0 Secret) (va_pbool_and (va_codegen_success_Mov128 (va_op_xmm_xmm\n 10) (va_op_xmm_xmm 0)) (va_pbool_and (va_codegen_success_Ghash_extra_bytes ()) (va_pbool_and\n (va_codegen_success_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_pbool_and\n (va_codegen_success_Pshufb (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (va_pbool_and\n (va_codegen_success_AESEncryptBlock alg) (va_pbool_and (va_codegen_success_Pxor (va_op_xmm_xmm\n 10) (va_op_xmm_xmm 0)) (va_pbool_and (va_codegen_success_Store128_buffer\n (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax) (va_op_xmm_xmm 10) 0 Secret)\n (va_ttrue ())))))))))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_extra_bytes (va_mods:va_mods_t) (alg:algorithm) (inout_b:buffer128) (key:(seq\n nat32)) (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (total_bytes:nat)\n (old_hash:quad32) (completed_quads:(seq quad32)) (h_LE:quad32) : (va_quickCode unit\n (va_code_Gcm_extra_bytes alg)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in let (len:(va_int_range\n 1 1)) = 1 in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 188 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64\n rRax) 0 Secret inout_b 0) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 189 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov128 (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (fun (va_s:va_state) _ -> let\n (hash_input:quad32) = va_get_xmm 0 va_s in va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 193 column 22 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Ghash_extra_bytes hkeys_b total_bytes old_hash h_LE completed_quads) (fun\n (va_s:va_state) _ -> va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 194 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (FStar.Seq.Base.equal #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s)\n inout_b) (FStar.Seq.Base.create #quad32 1 hash_input)) (let (snap:(FStar.Seq.Base.seq\n Vale.X64.Decls.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s) inout_b in va_QSeq\n va_range1\n \"***** PRECONDITION NOT MET AT line 198 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 199 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Pshufb (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (fun (va_s:va_state) _ -> va_QBind\n va_range1\n \"***** PRECONDITION NOT MET AT line 200 column 20 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_AESEncryptBlock alg (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 11 va_s)) key\n round_keys keys_b) (fun (va_s:va_state) _ -> va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 201 column 26 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.AES_s.aes_encrypt_LE_reveal ()) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 204 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Pxor (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 205 column 20 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Store128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax)\n (va_op_xmm_xmm 10) 0 Secret inout_b 0) (fun (va_s:va_state) _ -> va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 207 column 24 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.gctr_partial_reveal ()) (va_QEmpty (()))))))))))))))\n\n\nval va_lemma_Gcm_extra_bytes : va_b0:va_code -> va_s0:va_state -> alg:algorithm ->\n inout_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128 -> total_bytes:nat -> old_hash:quad32 -> completed_quads:(seq quad32) ->\n h_LE:quad32\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_extra_bytes alg) va_s0 /\\ va_get_ok va_s0 /\\ (let\n (len:(va_int_range 1 1)) = 1 in sse_enabled /\\ Vale.X64.Decls.buffers_disjoint128 keys_b\n inout_b /\\ Vale.X64.Decls.buffers_disjoint128 hkeys_b inout_b /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) (va_get_reg64 rRax va_s0) inout_b\n len (va_get_mem_layout va_s0) Secret /\\ len == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 inout_b /\\ va_get_xmm 9 va_s0 == Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\ aes_reqs alg key round_keys\n keys_b (va_get_reg64 rR8 va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\\n pclmulqdq_enabled /\\ Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0\n va_s0) hkeys_b) (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret /\\ va_get_xmm 8 va_s0 == Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.AES.GHash.ghash_incremental0 h_LE old_hash completed_quads) /\\ FStar.Seq.Base.length\n #quad32 completed_quads == total_bytes `op_Division` 16 /\\ total_bytes < 16 `op_Multiply`\n FStar.Seq.Base.length #quad32 completed_quads + 16 /\\ va_get_reg64 rR10 va_s0 == total_bytes\n `op_Modulus` 16 /\\ total_bytes `op_Modulus` 16 =!= 0 /\\ (0 < total_bytes /\\ total_bytes < 16\n `op_Multiply` Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes) /\\ 16 `op_Multiply`\n (Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes - 1) < total_bytes)))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (let (len:(va_int_range 1 1)) = 1 in Vale.X64.Decls.modifies_buffer128 inout_b\n (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5 va_sM) /\\ Vale.AES.GCTR.gctr_partial alg len\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b) key (va_get_xmm 11 va_s0) /\\ (let raw_quads =\n FStar.Seq.Base.append #quad32 completed_quads (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0)\n inout_b) in let input_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads) 0 total_bytes in let padded_bytes =\n Vale.AES.GCTR_s.pad_to_128_bits input_bytes in let input_quads =\n Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in l_and (FStar.Seq.Base.length\n #Vale.Def.Types_s.quad32 input_quads > 0) (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8\n va_sM) == Vale.AES.GHash.ghash_incremental h_LE old_hash input_quads))) /\\ va_state_eq va_sM\n (va_update_flags va_sM (va_update_mem_heaplet 5 va_sM (va_update_xmm 10 va_sM (va_update_xmm 8\n va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5 va_sM (va_update_xmm 4\n va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1 va_sM (va_update_xmm 0\n va_sM (va_update_reg64 rR11 va_sM (va_update_reg64 rRcx va_sM (va_update_ok va_sM\n (va_update_mem va_sM va_s0))))))))))))))))))\n[@\"opaque_to_smt\"]\nlet va_lemma_Gcm_extra_bytes va_b0 va_s0 alg inout_b key round_keys keys_b hkeys_b total_bytes\n old_hash completed_quads h_LE =\n let (va_mods:va_mods_t) = [va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8;\n va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm\n 1; va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_ok; va_Mod_mem] in\n let va_qc = va_qcode_Gcm_extra_bytes va_mods alg inout_b key round_keys keys_b hkeys_b\n total_bytes old_hash completed_quads h_LE in\n let (va_sM, va_fM, va_g) = va_wp_sound_code_norm (va_code_Gcm_extra_bytes alg) va_qc va_s0 (fun\n va_s0 va_sM va_g -> let () = va_g in label va_range1\n \"***** POSTCONDITION NOT MET AT line 121 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\ (let (len:(va_int_range 1 1)) = 1 in label va_range1\n \"***** POSTCONDITION NOT MET AT line 174 column 55 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 177 column 95 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.AES.GCTR.gctr_partial alg len (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM) inout_b) key (va_get_xmm 11 va_s0)) /\\ label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 180 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let raw_quads = FStar.Seq.Base.append #quad32 completed_quads (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 181 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let input_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads) 0 total_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 182 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let padded_bytes = Vale.AES.GCTR_s.pad_to_128_bits input_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 183 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let input_quads = Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 186 column 59 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (l_and (FStar.Seq.Base.length #Vale.Def.Types_s.quad32 input_quads > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE old_hash input_quads)))))))) in\n assert_norm (va_qc.mods == va_mods);\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm\n 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1;\n va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_ok; va_Mod_mem]) va_sM va_s0;\n (va_sM, va_fM)\n\n[@ va_qattr]\nlet va_wp_Gcm_extra_bytes (alg:algorithm) (inout_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (total_bytes:nat) (old_hash:quad32)\n (completed_quads:(seq quad32)) (h_LE:quad32) (va_s0:va_state) (va_k:(va_state -> unit -> Type0))\n : Type0 =\n (va_get_ok va_s0 /\\ (let (len:(va_int_range 1 1)) = 1 in sse_enabled /\\\n Vale.X64.Decls.buffers_disjoint128 keys_b inout_b /\\ Vale.X64.Decls.buffers_disjoint128 hkeys_b\n inout_b /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) (va_get_reg64 rRax\n va_s0) inout_b len (va_get_mem_layout va_s0) Secret /\\ len == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 inout_b /\\ va_get_xmm 9 va_s0 == Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\ aes_reqs alg key round_keys\n keys_b (va_get_reg64 rR8 va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\\n pclmulqdq_enabled /\\ Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0\n va_s0) hkeys_b) (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret /\\ va_get_xmm 8 va_s0 == Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.AES.GHash.ghash_incremental0 h_LE old_hash completed_quads) /\\ FStar.Seq.Base.length\n #quad32 completed_quads == total_bytes `op_Division` 16 /\\ total_bytes < 16 `op_Multiply`\n FStar.Seq.Base.length #quad32 completed_quads + 16 /\\ va_get_reg64 rR10 va_s0 == total_bytes\n `op_Modulus` 16 /\\ total_bytes `op_Modulus` 16 =!= 0 /\\ (0 < total_bytes /\\ total_bytes < 16\n `op_Multiply` Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes) /\\ 16 `op_Multiply`\n (Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes - 1) < total_bytes) /\\ (forall\n (va_x_mem:vale_heap) (va_x_rcx:nat64) (va_x_r11:nat64) (va_x_xmm0:quad32) (va_x_xmm1:quad32)\n (va_x_xmm2:quad32) (va_x_xmm3:quad32) (va_x_xmm4:quad32) (va_x_xmm5:quad32) (va_x_xmm6:quad32)\n (va_x_xmm7:quad32) (va_x_xmm8:quad32) (va_x_xmm10:quad32) (va_x_heap5:vale_heap)\n (va_x_efl:Vale.X64.Flags.t) . let va_sM = va_upd_flags va_x_efl (va_upd_mem_heaplet 5\n va_x_heap5 (va_upd_xmm 10 va_x_xmm10 (va_upd_xmm 8 va_x_xmm8 (va_upd_xmm 7 va_x_xmm7\n (va_upd_xmm 6 va_x_xmm6 (va_upd_xmm 5 va_x_xmm5 (va_upd_xmm 4 va_x_xmm4 (va_upd_xmm 3 va_x_xmm3\n (va_upd_xmm 2 va_x_xmm2 (va_upd_xmm 1 va_x_xmm1 (va_upd_xmm 0 va_x_xmm0 (va_upd_reg64 rR11\n va_x_r11 (va_upd_reg64 rRcx va_x_rcx (va_upd_mem va_x_mem va_s0)))))))))))))) in va_get_ok\n va_sM /\\ (let (len:(va_int_range 1 1)) = 1 in Vale.X64.Decls.modifies_buffer128 inout_b\n (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5 va_sM) /\\ Vale.AES.GCTR.gctr_partial alg len\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b) key (va_get_xmm 11 va_s0) /\\ (let raw_quads =\n FStar.Seq.Base.append #quad32 completed_quads (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0)\n inout_b) in let input_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads) 0 total_bytes in let padded_bytes =\n Vale.AES.GCTR_s.pad_to_128_bits input_bytes in let input_quads =\n Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in l_and (FStar.Seq.Base.length\n #Vale.Def.Types_s.quad32 input_quads > 0) (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8\n va_sM) == Vale.AES.GHash.ghash_incremental h_LE old_hash input_quads))) ==> va_k va_sM (())))\n\n\nval va_wpProof_Gcm_extra_bytes : alg:algorithm -> inout_b:buffer128 -> key:(seq nat32) ->\n round_keys:(seq quad32) -> keys_b:buffer128 -> hkeys_b:buffer128 -> total_bytes:nat ->\n old_hash:quad32 -> completed_quads:(seq quad32) -> h_LE:quad32 -> va_s0:va_state ->\n va_k:(va_state -> unit -> Type0)\n -> Ghost (va_state & va_fuel & unit)\n (requires (va_t_require va_s0 /\\ va_wp_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b\n total_bytes old_hash completed_quads h_LE va_s0 va_k))\n (ensures (fun (va_sM, va_f0, va_g) -> va_t_ensure (va_code_Gcm_extra_bytes alg) ([va_Mod_flags;\n va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5;\n va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR11;\n va_Mod_reg64 rRcx; va_Mod_mem]) va_s0 va_k ((va_sM, va_f0, va_g))))\n[@\"opaque_to_smt\"]\nlet va_wpProof_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b total_bytes old_hash\n completed_quads h_LE va_s0 va_k =\n let (va_sM, va_f0) = va_lemma_Gcm_extra_bytes (va_code_Gcm_extra_bytes alg) va_s0 alg inout_b key\n round_keys keys_b hkeys_b total_bytes old_hash completed_quads h_LE in\n va_lemma_upd_update va_sM;\n assert (va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 5 va_sM (va_update_xmm 10\n va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5\n va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1\n va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR11 va_sM (va_update_reg64 rRcx va_sM\n (va_update_ok va_sM (va_update_mem va_sM va_s0)))))))))))))))));\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm\n 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1;\n va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_mem]) va_sM va_s0;\n let va_g = () in\n (va_sM, va_f0, va_g)\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_quick_Gcm_extra_bytes (alg:algorithm) (inout_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (total_bytes:nat) (old_hash:quad32)\n (completed_quads:(seq quad32)) (h_LE:quad32) : (va_quickCode unit (va_code_Gcm_extra_bytes alg)) =\n (va_QProc (va_code_Gcm_extra_bytes alg) ([va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10;\n va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm\n 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_mem])\n (va_wp_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b total_bytes old_hash\n completed_quads h_LE) (va_wpProof_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b\n total_bytes old_hash completed_quads h_LE))\n//--\n//-- Gcm_blocks128\n\nval va_code_Gcm_blocks128 : alg:algorithm -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_blocks128 alg =\n (va_Block (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRbx) (va_op_opr64_reg64 rRdi))\n (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR12) (va_op_opr64_reg64 rRdx)) (va_CCons\n (va_code_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRax)) (va_CCons\n (va_code_Ghash_buffer ()) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRdi)\n (va_op_opr64_reg64 rRbx)) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRdx)\n (va_op_opr64_reg64 rR12)) (va_CCons (va_code_Gctr_blocks128 alg) (va_CNil ())))))))))\n\nval va_codegen_success_Gcm_blocks128 : alg:algorithm -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_blocks128 alg =\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rRbx) (va_op_opr64_reg64 rRdi))\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR12) (va_op_opr64_reg64 rRdx))\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRax))\n (va_pbool_and (va_codegen_success_Ghash_buffer ()) (va_pbool_and (va_codegen_success_Mov64\n (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRbx)) (va_pbool_and (va_codegen_success_Mov64\n (va_op_dst_opr64_reg64 rRdx) (va_op_opr64_reg64 rR12)) (va_pbool_and\n (va_codegen_success_Gctr_blocks128 alg) (va_ttrue ()))))))))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_blocks128 (va_mods:va_mods_t) (alg:algorithm) (in_b:buffer128) (out_b:buffer128)\n (key:(seq nat32)) (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (h_LE:quad32)\n : (va_quickCode unit (va_code_Gcm_blocks128 alg)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 274 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRbx) (va_op_opr64_reg64 rRdi)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 275 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR12) (va_op_opr64_reg64 rRdx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 276 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRax)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 277 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Ghash_buffer hkeys_b in_b h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8\n va_old_s))) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 278 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRbx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 279 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRdx) (va_op_opr64_reg64 rR12)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 280 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gctr_blocks128 alg in_b out_b key round_keys keys_b) (va_QEmpty (()))))))))))\n\n\nval va_lemma_Gcm_blocks128 : va_b0:va_code -> va_s0:va_state -> alg:algorithm -> in_b:buffer128 ->\n out_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128 -> h_LE:quad32\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_blocks128 alg) va_s0 /\\ va_get_ok va_s0 /\\\n (sse_enabled /\\ Vale.X64.Decls.buffers_disjoint128 keys_b out_b /\\\n Vale.X64.Decls.buffers_disjoint128 hkeys_b out_b /\\ (Vale.X64.Decls.buffers_disjoint128 in_b\n out_b \\/ in_b == out_b) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0)\n (va_get_reg64 rRax va_s0) in_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1 va_s0) (va_get_reg64 rRdi va_s0) out_b\n (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\ va_get_reg64 rRax va_s0 + 16\n `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply`\n va_get_reg64 rRdx va_s0 < pow2_64 /\\ l_and (Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 in_b == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out_b)\n (Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b < pow2_32) /\\ va_get_reg64 rRdx\n va_s0 == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b /\\ va_get_xmm 9 va_s0 ==\n Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\\n va_get_reg64 rRdx va_s0 < pow2_32 /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rR8\n va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret)))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (Vale.X64.Decls.modifies_buffer128 out_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM) /\\ Vale.AES.GCTR.gctr_partial alg (va_get_reg64 rRdx va_s0) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b) key\n (va_get_xmm 11 va_s0) /\\ va_get_xmm 11 va_sM == Vale.AES.GCTR.inc32lite (va_get_xmm 11 va_s0)\n (va_get_reg64 rRdx va_s0) /\\ (va_get_reg64 rRdx va_s0 == 0 ==> l_and (va_get_xmm 8 va_sM ==\n va_get_xmm 8 va_s0) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b ==\n Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) out_b)) /\\ (va_get_reg64 rRdx va_s0 > 0 ==>\n l_and (va_get_reg64 rRdx va_s0 <= FStar.Seq.Base.length #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) ==> FStar.Seq.Base.length\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.slice #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) 0 (va_get_reg64 rRdx va_s0)) > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s0)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b)))) /\\ va_state_eq va_sM (va_update_flags va_sM\n (va_update_mem_heaplet 1 va_sM (va_update_xmm 10 va_sM (va_update_xmm 11 va_sM (va_update_xmm 8\n va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5 va_sM (va_update_xmm 4\n va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1 va_sM (va_update_xmm 0\n va_sM (va_update_reg64 rR12 va_sM (va_update_reg64 rRdx va_sM (va_update_reg64 rR10 va_sM\n (va_update_reg64 rR11 va_sM (va_update_reg64 rRdi va_sM (va_update_reg64 rRbx va_sM\n (va_update_ok va_sM (va_update_mem va_sM va_s0)))))))))))))))))))))))\n[@\"opaque_to_smt\"]\nlet va_lemma_Gcm_blocks128 va_b0 va_s0 alg in_b out_b key round_keys keys_b hkeys_b h_LE =\n let (va_mods:va_mods_t) = [va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11;\n va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm\n 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10;\n va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_ok; va_Mod_mem] in\n let va_qc = va_qcode_Gcm_blocks128 va_mods alg in_b out_b key round_keys keys_b hkeys_b h_LE in\n let (va_sM, va_fM, va_g) = va_wp_sound_code_norm (va_code_Gcm_blocks128 alg) va_qc va_s0 (fun\n va_s0 va_sM va_g -> let () = va_g in label va_range1\n \"***** POSTCONDITION NOT MET AT line 210 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\ (label va_range1\n \"***** POSTCONDITION NOT MET AT line 255 column 53 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 261 column 95 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.AES.GCTR.gctr_partial alg (va_get_reg64 rRdx va_s0) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b) key\n (va_get_xmm 11 va_s0)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 262 column 45 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_xmm 11 va_sM == Vale.AES.GCTR.inc32lite (va_get_xmm 11 va_s0) (va_get_reg64 rRdx\n va_s0)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 265 column 93 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRdx va_s0 == 0 ==> l_and (va_get_xmm 8 va_sM == va_get_xmm 8 va_s0)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b == Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) out_b)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 267 column 131 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRdx va_s0 > 0 ==> l_and (va_get_reg64 rRdx va_s0 <= FStar.Seq.Base.length\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) ==>\n FStar.Seq.Base.length #Vale.X64.Decls.quad32 (FStar.Seq.Base.slice #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) 0 (va_get_reg64 rRdx va_s0)) > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s0)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b))))) in\n assert_norm (va_qc.mods == va_mods);\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11; va_Mod_xmm\n 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2;\n va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10;\n va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_ok; va_Mod_mem]) va_sM va_s0;\n (va_sM, va_fM)\n\n[@ va_qattr]\nlet va_wp_Gcm_blocks128 (alg:algorithm) (in_b:buffer128) (out_b:buffer128) (key:(seq nat32))\n (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (h_LE:quad32) (va_s0:va_state)\n (va_k:(va_state -> unit -> Type0)) : Type0 =\n (va_get_ok va_s0 /\\ (sse_enabled /\\ Vale.X64.Decls.buffers_disjoint128 keys_b out_b /\\\n Vale.X64.Decls.buffers_disjoint128 hkeys_b out_b /\\ (Vale.X64.Decls.buffers_disjoint128 in_b\n out_b \\/ in_b == out_b) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0)\n (va_get_reg64 rRax va_s0) in_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1 va_s0) (va_get_reg64 rRdi va_s0) out_b\n (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\ va_get_reg64 rRax va_s0 + 16\n `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply`\n va_get_reg64 rRdx va_s0 < pow2_64 /\\ l_and (Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 in_b == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out_b)\n (Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b < pow2_32) /\\ va_get_reg64 rRdx\n va_s0 == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b /\\ va_get_xmm 9 va_s0 ==\n Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\\n va_get_reg64 rRdx va_s0 < pow2_32 /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rR8\n va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret) /\\ (forall (va_x_mem:vale_heap) (va_x_rbx:nat64) (va_x_rdi:nat64) (va_x_r11:nat64)\n (va_x_r10:nat64) (va_x_rdx:nat64) (va_x_r12:nat64) (va_x_xmm0:quad32) (va_x_xmm1:quad32)\n (va_x_xmm2:quad32) (va_x_xmm3:quad32) (va_x_xmm4:quad32) (va_x_xmm5:quad32) (va_x_xmm6:quad32)\n (va_x_xmm7:quad32) (va_x_xmm8:quad32) (va_x_xmm11:quad32) (va_x_xmm10:quad32)\n (va_x_heap1:vale_heap) (va_x_efl:Vale.X64.Flags.t) . let va_sM = va_upd_flags va_x_efl\n (va_upd_mem_heaplet 1 va_x_heap1 (va_upd_xmm 10 va_x_xmm10 (va_upd_xmm 11 va_x_xmm11\n (va_upd_xmm 8 va_x_xmm8 (va_upd_xmm 7 va_x_xmm7 (va_upd_xmm 6 va_x_xmm6 (va_upd_xmm 5 va_x_xmm5\n (va_upd_xmm 4 va_x_xmm4 (va_upd_xmm 3 va_x_xmm3 (va_upd_xmm 2 va_x_xmm2 (va_upd_xmm 1 va_x_xmm1\n (va_upd_xmm 0 va_x_xmm0 (va_upd_reg64 rR12 va_x_r12 (va_upd_reg64 rRdx va_x_rdx (va_upd_reg64\n rR10 va_x_r10 (va_upd_reg64 rR11 va_x_r11 (va_upd_reg64 rRdi va_x_rdi (va_upd_reg64 rRbx\n va_x_rbx (va_upd_mem va_x_mem va_s0))))))))))))))))))) in va_get_ok va_sM /\\\n (Vale.X64.Decls.modifies_buffer128 out_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM) /\\ Vale.AES.GCTR.gctr_partial alg (va_get_reg64 rRdx va_s0) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b) key\n (va_get_xmm 11 va_s0) /\\ va_get_xmm 11 va_sM == Vale.AES.GCTR.inc32lite (va_get_xmm 11 va_s0)\n (va_get_reg64 rRdx va_s0) /\\ (va_get_reg64 rRdx va_s0 == 0 ==> l_and (va_get_xmm 8 va_sM ==\n va_get_xmm 8 va_s0) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b ==\n Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) out_b)) /\\ (va_get_reg64 rRdx va_s0 > 0 ==>\n l_and (va_get_reg64 rRdx va_s0 <= FStar.Seq.Base.length #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) ==> FStar.Seq.Base.length\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.slice #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) 0 (va_get_reg64 rRdx va_s0)) > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s0)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b)))) ==> va_k va_sM (())))\n\n\nval va_wpProof_Gcm_blocks128 : alg:algorithm -> in_b:buffer128 -> out_b:buffer128 -> key:(seq\n nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 -> hkeys_b:buffer128 -> h_LE:quad32 ->\n va_s0:va_state -> va_k:(va_state -> unit -> Type0)\n -> Ghost (va_state & va_fuel & unit)\n (requires (va_t_require va_s0 /\\ va_wp_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b\n h_LE va_s0 va_k))\n (ensures (fun (va_sM, va_f0, va_g) -> va_t_ensure (va_code_Gcm_blocks128 alg) ([va_Mod_flags;\n va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6;\n va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0;\n va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10; va_Mod_reg64 rR11; va_Mod_reg64 rRdi;\n va_Mod_reg64 rRbx; va_Mod_mem]) va_s0 va_k ((va_sM, va_f0, va_g))))\n[@\"opaque_to_smt\"]\nlet va_wpProof_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b h_LE va_s0 va_k =\n let (va_sM, va_f0) = va_lemma_Gcm_blocks128 (va_code_Gcm_blocks128 alg) va_s0 alg in_b out_b key\n round_keys keys_b hkeys_b h_LE in\n va_lemma_upd_update va_sM;\n assert (va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 1 va_sM (va_update_xmm 10\n va_sM (va_update_xmm 11 va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6\n va_sM (va_update_xmm 5 va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2\n va_sM (va_update_xmm 1 va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR12 va_sM\n (va_update_reg64 rRdx va_sM (va_update_reg64 rR10 va_sM (va_update_reg64 rR11 va_sM\n (va_update_reg64 rRdi va_sM (va_update_reg64 rRbx va_sM (va_update_ok va_sM (va_update_mem\n va_sM va_s0))))))))))))))))))))));\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11; va_Mod_xmm\n 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2;\n va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10;\n va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_mem]) va_sM va_s0;\n let va_g = () in\n (va_sM, va_f0, va_g)\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_quick_Gcm_blocks128 (alg:algorithm) (in_b:buffer128) (out_b:buffer128) (key:(seq nat32))\n (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (h_LE:quad32) : (va_quickCode\n unit (va_code_Gcm_blocks128 alg)) =\n (va_QProc (va_code_Gcm_blocks128 alg) ([va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10;\n va_Mod_xmm 11; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm\n 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64\n rR10; va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_mem])\n (va_wp_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b h_LE)\n (va_wpProof_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b h_LE))\n//--\n//-- Gcm_blocks\n#push-options \"--z3rlimit 1000\"\n\nval va_code_Gcm_blocks : alg:algorithm -> offset:int -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_blocks alg offset =\n (va_Block (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR13) (va_op_opr64_reg64 rRcx))\n (va_CCons (va_code_AddLea64 (va_op_dst_opr64_reg64 rR9) (va_op_opr64_reg64 rR9) (va_const_opr64\n 32)) (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rRbx) (va_op_reg_opr64_reg64 rRsp)\n (offset + 0)) (va_CCons (va_code_Gcm_blocks_auth ()) (va_CCons (va_code_Load64_stack\n (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 8)) (va_CCons\n (va_code_Load64_stack (va_op_dst_opr64_reg64 rRsi) (va_op_reg_opr64_reg64 rRsp) (offset + 16))\n (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp)\n (offset + 24)) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRcx) (va_op_opr64_reg64 rR13))\n (va_CCons (va_code_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (va_CCons\n (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 2) (va_op_xmm_xmm 1) (va_op_reg_opr64_reg64\n rR8) 0 Public) (va_CCons (va_code_Store128_buffer (va_op_heaplet_mem_heaplet 3)\n (va_op_reg_opr64_reg64 rRbp) (va_op_xmm_xmm 1) 0 Secret) (va_CCons (va_code_Load_one_lsb\n (va_op_xmm_xmm 10)) (va_CCons (va_code_VPaddd (va_op_xmm_xmm 1) (va_op_xmm_xmm 1)\n (va_op_xmm_xmm 10)) (va_CCons (va_code_AES_GCM_decrypt_6mult alg) (va_CCons\n (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 11) (va_op_reg_opr64_reg64\n rRbp) 32 Secret) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR8) (va_op_opr64_reg64 rRcx))\n (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp)\n (offset + 32)) (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rRdi)\n (va_op_reg_opr64_reg64 rRsp) (offset + 40)) (va_CCons (va_code_Load64_stack\n (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp) (offset + 48)) (va_CCons\n (va_code_Mov64 (va_op_dst_opr64_reg64 rR14) (va_op_opr64_reg64 rRdx)) (va_CCons\n (va_code_InitPshufbMask (va_op_xmm_xmm 9) (va_op_reg_opr64_reg64 rR12)) (va_CCons\n (va_code_Pshufb (va_op_xmm_xmm 11) (va_op_xmm_xmm 9)) (va_CCons (va_code_Gcm_blocks128 alg)\n (va_CCons (va_code_Stack_lemma ()) (va_CCons (va_code_Add64 (va_op_dst_opr64_reg64 rR14)\n (va_opr_code_Stack (va_op_reg64_reg64 rRsp) (offset + 24) Public)) (va_CCons (va_code_IMul64\n (va_op_dst_opr64_reg64 rR14) (va_const_opr64 16)) (va_CCons (va_code_Load64_stack\n (va_op_dst_opr64_reg64 rR13) (va_op_reg_opr64_reg64 rRsp) (offset + 64)) (va_CCons (va_IfElse\n (va_cmp_gt (va_op_cmp_reg64 rR13) (va_op_cmp_reg64 rR14)) (va_Block (va_CCons\n (va_code_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp) (offset + 56))\n (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR10) (va_op_opr64_reg64 rR13)) (va_CCons\n (va_code_And64 (va_op_dst_opr64_reg64 rR10) (va_const_opr64 15)) (va_CCons\n (va_code_Gcm_extra_bytes alg) (va_CCons (va_Block (va_CNil ())) (va_CNil ()))))))) (va_Block\n (va_CNil ()))) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR11) (va_op_opr64_reg64 rR15))\n (va_CCons (va_code_Gcm_make_length_quad ()) (va_CCons (va_code_Ghash_register ()) (va_CCons\n (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64\n rRbp) 0 Secret) (va_CCons (va_code_Gctr_register alg) (va_CCons (va_Block (va_CNil ()))\n (va_CNil ()))))))))))))))))))))))))))))))))))))\n\nval va_codegen_success_Gcm_blocks : alg:algorithm -> offset:int -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_blocks alg offset =\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR13) (va_op_opr64_reg64 rRcx))\n (va_pbool_and (va_codegen_success_AddLea64 (va_op_dst_opr64_reg64 rR9) (va_op_opr64_reg64 rR9)\n (va_const_opr64 32)) (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64\n rRbx) (va_op_reg_opr64_reg64 rRsp) (offset + 0)) (va_pbool_and\n (va_codegen_success_Gcm_blocks_auth ()) (va_pbool_and (va_codegen_success_Load64_stack\n (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 8)) (va_pbool_and\n (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRsi) (va_op_reg_opr64_reg64 rRsp)\n (offset + 16)) (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRdx)\n (va_op_reg_opr64_reg64 rRsp) (offset + 24)) (va_pbool_and (va_codegen_success_Mov64\n (va_op_dst_opr64_reg64 rRcx) (va_op_opr64_reg64 rR13)) (va_pbool_and (va_codegen_success_Mov128\n (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (va_pbool_and (va_codegen_success_Load128_buffer\n (va_op_heaplet_mem_heaplet 2) (va_op_xmm_xmm 1) (va_op_reg_opr64_reg64 rR8) 0 Public)\n (va_pbool_and (va_codegen_success_Store128_buffer (va_op_heaplet_mem_heaplet 3)\n (va_op_reg_opr64_reg64 rRbp) (va_op_xmm_xmm 1) 0 Secret) (va_pbool_and\n (va_codegen_success_Load_one_lsb (va_op_xmm_xmm 10)) (va_pbool_and (va_codegen_success_VPaddd\n (va_op_xmm_xmm 1) (va_op_xmm_xmm 1) (va_op_xmm_xmm 10)) (va_pbool_and\n (va_codegen_success_AES_GCM_decrypt_6mult alg) (va_pbool_and (va_codegen_success_Load128_buffer\n (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 11) (va_op_reg_opr64_reg64 rRbp) 32 Secret)\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR8) (va_op_opr64_reg64 rRcx))\n (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRax)\n (va_op_reg_opr64_reg64 rRsp) (offset + 32)) (va_pbool_and (va_codegen_success_Load64_stack\n (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 40)) (va_pbool_and\n (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp)\n (offset + 48)) (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR14)\n (va_op_opr64_reg64 rRdx)) (va_pbool_and (va_codegen_success_InitPshufbMask (va_op_xmm_xmm 9)\n (va_op_reg_opr64_reg64 rR12)) (va_pbool_and (va_codegen_success_Pshufb (va_op_xmm_xmm 11)\n (va_op_xmm_xmm 9)) (va_pbool_and (va_codegen_success_Gcm_blocks128 alg) (va_pbool_and\n (va_codegen_success_Stack_lemma ()) (va_pbool_and (va_codegen_success_Add64\n (va_op_dst_opr64_reg64 rR14) (va_opr_code_Stack (va_op_reg64_reg64 rRsp) (offset + 24) Public))\n (va_pbool_and (va_codegen_success_IMul64 (va_op_dst_opr64_reg64 rR14) (va_const_opr64 16))\n (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rR13)\n (va_op_reg_opr64_reg64 rRsp) (offset + 64)) (va_pbool_and (va_pbool_and\n (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp)\n (offset + 56)) (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR10)\n (va_op_opr64_reg64 rR13)) (va_pbool_and (va_codegen_success_And64 (va_op_dst_opr64_reg64 rR10)\n (va_const_opr64 15)) (va_codegen_success_Gcm_extra_bytes alg)))) (va_pbool_and\n (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR11) (va_op_opr64_reg64 rR15)) (va_pbool_and\n (va_codegen_success_Gcm_make_length_quad ()) (va_pbool_and (va_codegen_success_Ghash_register\n ()) (va_pbool_and (va_codegen_success_Load128_buffer (va_op_heaplet_mem_heaplet 3)\n (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64 rRbp) 0 Secret) (va_pbool_and\n (va_codegen_success_Gctr_register alg) (va_ttrue ()))))))))))))))))))))))))))))))))))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_blocks (va_mods:va_mods_t) (alg:algorithm) (offset:int) (auth_b:buffer128)\n (abytes_b:buffer128) (in128x6_b:buffer128) (out128x6_b:buffer128) (in128_b:buffer128)\n (out128_b:buffer128) (inout_b:buffer128) (iv_b:buffer128) (scratch_b:buffer128) (key:(seq nat32))\n (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) : (va_quickCode unit\n (va_code_Gcm_blocks alg offset)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in let\n (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 0) (va_get_stack va_s) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 8) (va_get_stack va_s) in let\n (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 16) (va_get_stack va_s) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 24) (va_get_stack va_s) in let\n (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 32) (va_get_stack va_s) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 40) (va_get_stack va_s) in let\n (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset\n + 48) (va_get_stack va_s) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 56) (va_get_stack va_s) in let\n (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s\n + offset + 64) (va_get_stack va_s) in let (h_LE:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.X64.Decls.buffer128_read hkeys_b 2\n (va_get_mem_heaplet 0 va_old_s)) in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 463 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR13) (va_op_opr64_reg64 rRcx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 464 column 13 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_AddLea64 (va_op_dst_opr64_reg64 rR9) (va_op_opr64_reg64 rR9) (va_const_opr64 32))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 465 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRbx) (va_op_reg_opr64_reg64 rRsp) (offset + 0))\n (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 466 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_blocks_auth auth_b abytes_b hkeys_b h_LE) (fun (va_s:va_state)\n (auth_quad_seq:(seq quad32)) -> let (y_0:quad32) = Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 0 0 0 0 in let (y_auth_bytes:quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s) in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 473 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 8))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 474 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRsi) (va_op_reg_opr64_reg64 rRsp) (offset + 16))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 475 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp) (offset + 24))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 476 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRcx) (va_op_opr64_reg64 rR13)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 477 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (fun (va_s:va_state) _ -> let\n (iv_BE:Vale.X64.Decls.quad32) = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2\n va_old_s) in let (ctr_BE_1:quad32) = iv_BE in let (ctr_BE_2:quad32) = Vale.AES.GCTR_s.inc32\n iv_BE 1 in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 483 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 2) (va_op_xmm_xmm 1) (va_op_reg_opr64_reg64\n rR8) 0 Public iv_b 0) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 485 column 20 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Store128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_reg_opr64_reg64 rRbp)\n (va_op_xmm_xmm 1) 0 Secret scratch_b 0) (fun (va_s:va_state) _ -> let (j0:quad32) = va_get_xmm\n 1 va_s in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 487 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load_one_lsb (va_op_xmm_xmm 10)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 489 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_VPaddd (va_op_xmm_xmm 1) (va_op_xmm_xmm 1) (va_op_xmm_xmm 10)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 491 column 26 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_AES_GCM_decrypt_6mult alg h_LE iv_b in128x6_b out128x6_b scratch_b key round_keys\n keys_b hkeys_b) (fun (va_s:va_state) _ -> let (y_cipher128x6:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s) in let (auth_in:(seq quad32)) =\n auth_quad_seq in let (va_arg138:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b in let\n (va_arg137:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = auth_in in let\n (va_arg136:Vale.Def.Types_s.quad32) = y_auth_bytes in let (va_arg135:Vale.Def.Types_s.quad32) =\n y_0 in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 494 column 36 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GHash.lemma_ghash_incremental0_append h_LE va_arg135 va_arg136\n y_cipher128x6 va_arg137 va_arg138) (let auth_in = FStar.Seq.Base.append #quad32 auth_in\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b) in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 498 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 11)\n (va_op_reg_opr64_reg64 rRbp) 32 Secret scratch_b 2) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 499 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR8) (va_op_opr64_reg64 rRcx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 500 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp) (offset + 32))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 501 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 40))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 502 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp) (offset + 48))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 503 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR14) (va_op_opr64_reg64 rRdx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 504 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_InitPshufbMask (va_op_xmm_xmm 9) (va_op_reg_opr64_reg64 rR12)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 505 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Pshufb (va_op_xmm_xmm 11) (va_op_xmm_xmm 9)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 506 column 18 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_blocks128 alg in128_b out128_b key round_keys keys_b hkeys_b h_LE) (fun\n (va_s:va_state) _ -> let (y_cipher128:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s) in let (va_arg134:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b in let\n (va_arg133:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = auth_in in let\n (va_arg132:Vale.Def.Types_s.quad32) = y_0 in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 508 column 36 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GHash.lemma_ghash_incremental0_append h_LE va_arg132 y_cipher128x6\n y_cipher128 va_arg133 va_arg134) (let auth_in = FStar.Seq.Base.append #quad32 auth_in\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b) in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 512 column 22 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Stack_lemma (va_op_reg64_reg64 rRsp) (offset + 24) Public) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 512 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Add64 (va_op_dst_opr64_reg64 rR14) (va_opr_code_Stack (va_op_reg64_reg64 rRsp)\n (offset + 24) Public)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 513 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_IMul64 (va_op_dst_opr64_reg64 rR14) (va_const_opr64 16)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 514 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rR13) (va_op_reg_opr64_reg64 rRsp) (offset + 64))\n (fun (va_s:va_state) _ -> let (y_inout:Vale.Def.Types_s.quad32) = y_cipher128 in let\n (plain_byte_seq:(seq quad32)) = empty_seq_quad32 in let (cipher_byte_seq:(seq quad32)) =\n empty_seq_quad32 in let (va_arg131:Vale.Def.Types_s.quad32) = va_get_xmm 11 va_s in let\n (va_arg130:(FStar.Seq.Base.seq Vale.Def.Types_s.nat32)) = key in let\n (va_arg129:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = cipher_byte_seq in let\n (va_arg128:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = plain_byte_seq in let\n (va_arg127:Vale.AES.AES_common_s.algorithm) = alg in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 519 column 29 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.gctr_partial_opaque_init va_arg127 va_arg128 va_arg129 va_arg130\n va_arg131) (let (total_bytes:(va_int_at_least 0)) = FStar.Seq.Base.length #quad32 auth_quad_seq\n `op_Multiply` 16 + plain_num_bytes in va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 523 column 8 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_qIf va_mods (Cmp_gt (va_op_cmp_reg64 rR13) (va_op_cmp_reg64 rR14)) (qblock va_mods (fun\n (va_s:va_state) -> va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 525 column 21 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp) (offset + 56))\n (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 526 column 14 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR10) (va_op_opr64_reg64 rR13)) (fun (va_s:va_state) _\n -> va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 527 column 26 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.Poly1305.Math.lemma_poly_bits64 ()) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 528 column 14 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_And64 (va_op_dst_opr64_reg64 rR10) (va_const_opr64 15)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 532 column 24 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b total_bytes y_0 auth_in\n h_LE) (fun (va_s:va_state) _ -> let y_inout = Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm\n 8 va_s) in let (raw_auth_quads:(FStar.Seq.Base.seq quad32)) = FStar.Seq.Base.append #quad32\n auth_in (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_old_s) inout_b) in va_qAssertSquash\n va_range1\n \"***** EXPRESSION PRECONDITIONS NOT MET WITHIN line 536 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n ((fun a_1906 (s_1907:(FStar.Seq.Base.seq a_1906)) (i_1908:Prims.nat) (j_1909:Prims.nat) -> let\n (j_1869:Prims.nat) = j_1909 in Prims.b2t (Prims.op_AmpAmp (Prims.op_LessThanOrEqual i_1908\n j_1869) (Prims.op_LessThanOrEqual j_1869 (FStar.Seq.Base.length #a_1906 s_1907))))\n Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_auth_quads) 0 total_bytes)\n (fun _ -> let (auth_input_bytes:(FStar.Seq.Base.seq Vale.Def.Types_s.nat8)) =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n raw_auth_quads) 0 total_bytes in let (padded_auth_bytes:(FStar.Seq.Base.seq\n Vale.Def.Types_s.nat8)) = Vale.AES.GCTR_s.pad_to_128_bits auth_input_bytes in let auth_in =\n Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_auth_bytes in let plain_byte_seq =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_old_s) inout_b in let cipher_byte_seq =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s) inout_b in va_QEmpty ((auth_in,\n cipher_byte_seq, plain_byte_seq, y_inout)))))))))) (qblock va_mods (fun (va_s:va_state) ->\n va_QEmpty ((auth_in, cipher_byte_seq, plain_byte_seq, y_inout))))) (fun (va_s:va_state) va_g ->\n let ((auth_in:(seq quad32)), (cipher_byte_seq:(seq quad32)), (plain_byte_seq:(seq quad32)),\n (y_inout:Vale.Def.Types_s.quad32)) = va_g in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 547 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR11) (va_op_opr64_reg64 rR15)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 548 column 25 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_make_length_quad ()) (fun (va_s:va_state) _ -> let\n (length_quad32:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 0\n va_s) in va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 551 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Ghash_register hkeys_b h_LE y_inout) (fun (va_s:va_state) _ -> let\n (y_final:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s)\n in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 554 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64\n rRbp) 0 Secret scratch_b 0) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 557 column 18 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gctr_register alg key round_keys keys_b) (fun (va_s:va_state) _ -> let\n (va_arg126:Vale.Def.Types_s.quad32) = va_get_xmm 8 va_s in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 560 column 40 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.Arch.Types.le_seq_quad32_to_bytes_of_singleton va_arg126)\n (va_qAssertSquash va_range1\n \"***** EXPRESSION PRECONDITIONS NOT MET WITHIN line 561 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n ((fun (icb_BE_677:Vale.Def.Types_s.quad32) (plain_LE_678:Vale.Def.Types_s.quad32)\n (alg_679:Vale.AES.AES_common_s.algorithm) (key_680:(FStar.Seq.Base.seq Vale.Def.Types_s.nat32))\n (i_681:Prims.int) -> Vale.AES.AES_s.is_aes_key_LE alg_679 key_680) j0 y_final alg key 0) (fun _\n -> va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 561 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_xmm 8 va_s == Vale.AES.GCTR_s.gctr_encrypt_block j0 y_final alg key 0) (let\n (plain128:(FStar.Seq.Base.seq Vale.X64.Decls.quad32)) = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b) in let\n (cipher128:(FStar.Seq.Base.seq Vale.X64.Decls.quad32)) = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) in128_b) in va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 566 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (FStar.Seq.Base.length #quad32 plain_byte_seq == 0 ==> FStar.Seq.Base.equal\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 plain128 plain_byte_seq)\n plain128) (va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 567 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (FStar.Seq.Base.length #quad32 cipher_byte_seq == 0 ==> FStar.Seq.Base.equal\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 cipher128 cipher_byte_seq)\n cipher128) (let (va_arg125:Vale.Def.Types_s.quad32) = Vale.AES.GCTR.inc32lite ctr_BE_2 len128x6\n in let (va_arg124:Vale.Def.Types_s.quad32) = ctr_BE_2 in let (va_arg123:(FStar.Seq.Base.seq\n Vale.Def.Types_s.nat32)) = key in let (va_arg122:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32))\n = Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) out128_b in let\n (va_arg121:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_old_s) in128_b in let (va_arg120:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s) out128x6_b in let\n (va_arg119:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_old_s) in128x6_b in let (va_arg118:Prims.nat) = len128 in let\n (va_arg117:Prims.nat) = len128x6 in let (va_arg116:Vale.AES.AES_common_s.algorithm) = alg in\n va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 569 column 30 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.lemma_gctr_partial_append va_arg116 va_arg117 va_arg118\n va_arg119 va_arg120 va_arg121 va_arg122 va_arg123 va_arg124 va_arg125) (let\n (va_arg115:Vale.Def.Types_s.quad32) = Vale.AES.GCTR.inc32lite (Vale.AES.GCTR.inc32lite ctr_BE_2\n len128x6) len128 in let (va_arg114:Vale.Def.Types_s.quad32) = ctr_BE_2 in let\n (va_arg113:(FStar.Seq.Base.seq Vale.Def.Types_s.nat32)) = key in let\n (va_arg112:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = cipher_byte_seq in let\n (va_arg111:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = plain_byte_seq in let\n (va_arg110:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s) out128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) out128_b) in let\n (va_arg109:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b) in let (va_arg108:Prims.nat) =\n FStar.Seq.Base.length #quad32 plain_byte_seq in let (va_arg107:Prims.nat) = len128x6 + len128\n in let (va_arg106:Vale.AES.AES_common_s.algorithm) = alg in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 575 column 30 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.lemma_gctr_partial_append va_arg106 va_arg107 va_arg108\n va_arg109 va_arg110 va_arg111 va_arg112 va_arg113 va_arg114 va_arg115) (let\n (va_arg105:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = auth_in in let\n (va_arg104:Vale.Def.Types_s.quad32) = y_0 in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 583 column 23 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GHash.lemma_hash_append2 h_LE va_arg104 y_inout y_final va_arg105\n length_quad32) (let auth_in = FStar.Seq.Base.append #quad32 auth_in (FStar.Seq.Base.create\n #Vale.Def.Types_s.quad32 1 length_quad32) in let (va_arg103:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = auth_in in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 585 column 31 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GHash.ghash_incremental_to_ghash h_LE va_arg103) (va_QEmpty\n (()))))))))))))))))))))))))))))))))))))))))))))))))\n\n\nval va_lemma_Gcm_blocks : va_b0:va_code -> va_s0:va_state -> alg:algorithm -> offset:int ->\n auth_b:buffer128 -> abytes_b:buffer128 -> in128x6_b:buffer128 -> out128x6_b:buffer128 ->\n in128_b:buffer128 -> out128_b:buffer128 -> inout_b:buffer128 -> iv_b:buffer128 ->\n scratch_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_blocks alg offset) va_s0 /\\ va_get_ok va_s0 /\\\n (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in let (h_LE:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.X64.Decls.buffer128_read hkeys_b 2\n (va_get_mem_heaplet 0 va_s0)) in sse_enabled /\\ movbe_enabled /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0)\n Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp\n va_s0 + offset + 8) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 24) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 40) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 56) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet\n 1 va_s0) (va_get_reg64 rRdi va_s0) auth_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0)\n Secret /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 7 va_s0) abytes_ptr abytes_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 2\n va_s0) (va_get_reg64 rR8 va_s0) iv_b 1 (va_get_mem_layout va_s0) Public /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 6 va_s0) in128x6_ptr in128x6_b len128x6\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 6\n va_s0) out128x6_ptr out128x6_b len128x6 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0) in128_ptr in128_b len128\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1\n va_s0) out128_ptr out128_b len128 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) inout_ptr inout_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 3\n va_s0) (va_get_reg64 rRbp va_s0) scratch_b 9 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0) hkeys_b 8\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.buffer_disjoints128 iv_b ([keys_b;\n scratch_b; in128x6_b; out128x6_b; hkeys_b; in128_b; out128_b; inout_b]) /\\\n Vale.X64.Decls.buffer_disjoints128 scratch_b ([keys_b; in128x6_b; out128x6_b; in128_b;\n out128_b; inout_b; hkeys_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128x6_b ([keys_b;\n hkeys_b; in128_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128_b ([keys_b; hkeys_b;\n out128x6_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 inout_b ([keys_b; hkeys_b;\n out128x6_b; out128_b]) /\\ (Vale.X64.Decls.buffers_disjoint128 in128x6_b out128x6_b \\/ in128x6_b\n == out128x6_b) /\\ (Vale.X64.Decls.buffers_disjoint128 in128_b out128_b \\/ in128_b == out128_b)\n /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ in128x6_ptr\n + 16 `op_Multiply` len128x6 < pow2_64 /\\ out128x6_ptr + 16 `op_Multiply` len128x6 < pow2_64 /\\\n in128_ptr + 16 `op_Multiply` len128 < pow2_64 /\\ out128_ptr + 16 `op_Multiply` len128 < pow2_64\n /\\ inout_ptr + 16 < pow2_64 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 auth_b ==\n va_get_reg64 rRdx va_s0 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 abytes_b == 1\n /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128x6_b ==\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out128x6_b /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 out128_b /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128\n in128x6_b == len128x6 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b ==\n len128 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 inout_b == 1 /\\\n plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32 /\\ va_get_reg64 rR9 va_s0 + 32 <\n pow2_64 /\\ Vale.X64.Memory.buffer_addr #Vale.X64.Memory.vuint128 keys_b (va_get_mem_heaplet 0\n va_s0) + 128 < pow2_64 /\\ len128x6 `op_Modulus` 6 == 0 /\\ (len128x6 > 0 ==> len128x6 >= 6) /\\\n 12 + len128x6 + 6 < pow2_32 /\\ (va_mul_nat len128x6 (128 `op_Division` 8) + va_mul_nat len128\n (128 `op_Division` 8) <= plain_num_bytes /\\ plain_num_bytes < va_mul_nat len128x6 (128\n `op_Division` 8) + va_mul_nat len128 (128 `op_Division` 8) + 128 `op_Division` 8) /\\\n (va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) <= va_get_reg64 rRsi va_s0 /\\\n va_get_reg64 rRsi va_s0 < va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) + 128\n `op_Division` 8) /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rRcx va_s0)\n (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 (Vale.AES.AES_s.aes_encrypt_LE alg key\n (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0))))))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in let (h_LE:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.X64.Decls.buffer128_read hkeys_b 2\n (va_get_mem_heaplet 0 va_s0)) in Vale.X64.Decls.modifies_buffer128 out128_b (va_get_mem_heaplet\n 1 va_s0) (va_get_mem_heaplet 1 va_sM) /\\ Vale.X64.Decls.modifies_buffer128 iv_b\n (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2 va_sM) /\\ Vale.X64.Decls.modifies_buffer128\n scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3 va_sM) /\\\n Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM) /\\ Vale.X64.Decls.modifies_buffer128 out128x6_b (va_get_mem_heaplet 6 va_s0)\n (va_get_mem_heaplet 6 va_sM) /\\ plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32\n /\\ (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in let\n (ctr_BE_1:quad32) = iv_BE in let (ctr_BE_2:quad32) = Vale.AES.GCTR_s.inc32 iv_BE 1 in let\n (plain_in:(seq quad32)) = (if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) then FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b) else FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in128_b)) in let (cipher_out:(seq quad32)) = (if (plain_num_bytes\n > (len128x6 + len128) `op_Multiply` 128 `op_Division` 8) then FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM)\n out128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM) inout_b) else\n FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM)\n out128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out128_b)) in let\n (cipher_bound:nat) = (if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128 `op_Division`\n 8) then (len128x6 + len128 + 1) else (len128x6 + len128)) in Vale.AES.GCTR.gctr_partial alg\n cipher_bound plain_in cipher_out key ctr_BE_2 /\\ (let (length_quad:quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.Def.Types_s.insert_nat64\n (Vale.Def.Types_s.insert_nat64 (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0) (8\n `op_Multiply` va_get_reg64 rRsi va_s0) 1) (8 `op_Multiply` plain_num_bytes) 0) in let\n (raw_auth_quads:(seq quad32)) = (if (va_get_reg64 rRsi va_s0 > va_get_reg64 rRdx va_s0\n `op_Multiply` 128 `op_Division` 8) then FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 7 va_s0) abytes_b) else Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0)\n auth_b) in let (auth_input_bytes:(seq nat8)) = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_auth_quads) 0 (va_get_reg64 rRsi va_s0) in let\n (padded_auth_bytes:(seq nat8)) = Vale.AES.GCTR_s.pad_to_128_bits auth_input_bytes in let\n (auth_quad_seq:(seq quad32)) = Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_auth_bytes in let\n (raw_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32 (FStar.Seq.Base.append #quad32\n auth_quad_seq (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b))\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b) in let (total_bytes:nat) =\n FStar.Seq.Base.length #quad32 auth_quad_seq `op_Multiply` 16 + plain_num_bytes in let\n (raw_quad_seq:(seq quad32)) = (if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) then (let (ab:(seq nat8)) = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes (FStar.Seq.Base.append #quad32 raw_quad_seq\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b))) 0 total_bytes in let (pb:(seq\n nat8)) = Vale.AES.GCTR_s.pad_to_128_bits ab in Vale.Def.Types_s.le_bytes_to_seq_quad32 pb) else\n raw_quad_seq) in let (auth_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32 raw_quad_seq\n (FStar.Seq.Base.create #quad32 1 length_quad) in va_get_xmm 8 va_sM ==\n Vale.AES.GCTR_s.gctr_encrypt_block ctr_BE_1 (Vale.AES.GHash_s.ghash_LE h_LE auth_quad_seq) alg\n key 0))) /\\ va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 6 va_sM\n (va_update_mem_heaplet 5 va_sM (va_update_mem_heaplet 3 va_sM (va_update_mem_heaplet 2 va_sM\n (va_update_mem_heaplet 1 va_sM (va_update_xmm 15 va_sM (va_update_xmm 14 va_sM (va_update_xmm\n 13 va_sM (va_update_xmm 12 va_sM (va_update_xmm 11 va_sM (va_update_xmm 10 va_sM (va_update_xmm\n 9 va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5\n va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1\n va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR15 va_sM (va_update_reg64 rR14 va_sM\n (va_update_reg64 rR13 va_sM (va_update_reg64 rR12 va_sM (va_update_reg64 rR11 va_sM\n (va_update_reg64 rR10 va_sM (va_update_reg64 rR9 va_sM (va_update_reg64 rR8 va_sM\n (va_update_reg64 rRbp va_sM (va_update_reg64 rRsi va_sM (va_update_reg64 rRdi va_sM\n (va_update_reg64 rRdx va_sM (va_update_reg64 rRcx va_sM (va_update_reg64 rRbx va_sM\n (va_update_reg64 rRax va_sM (va_update_ok va_sM (va_update_mem va_sM\n va_s0)))))))))))))))))))))))))))))))))))))))))\n[@\"opaque_to_smt\"]\nlet va_lemma_Gcm_blocks va_b0 va_s0 alg offset auth_b abytes_b in128x6_b out128x6_b in128_b\n out128_b inout_b iv_b scratch_b key round_keys keys_b hkeys_b =\n let (va_mods:va_mods_t) = [va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5;\n va_Mod_mem_heaplet 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14;\n va_Mod_xmm 13; va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8;\n va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm\n 1; va_Mod_xmm 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12;\n va_Mod_reg64 rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp;\n va_Mod_reg64 rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx;\n va_Mod_reg64 rRax; va_Mod_ok; va_Mod_mem] in\n let va_qc = va_qcode_Gcm_blocks va_mods alg offset auth_b abytes_b in128x6_b out128x6_b in128_b\n out128_b inout_b iv_b scratch_b key round_keys keys_b hkeys_b in\n let (va_sM, va_fM, va_g) = va_wp_sound_code_norm (va_code_Gcm_blocks alg offset) va_qc va_s0 (fun\n va_s0 va_sM va_g -> let () = va_g in label va_range1\n \"***** POSTCONDITION NOT MET AT line 283 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\ (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64\n (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0) in let\n (in128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 8) (va_get_stack va_s0) in let (out128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack va_s0) in\n let (len128x6:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 24) (va_get_stack va_s0) in let (in128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack va_s0) in\n let (out128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 40) (va_get_stack va_s0) in let (len128:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack va_s0) in\n let (inout_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 56) (va_get_stack va_s0) in let (plain_num_bytes:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack va_s0) in\n let (h_LE:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.X64.Decls.buffer128_read hkeys_b 2 (va_get_mem_heaplet 0 va_s0)) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 396 column 56 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out128_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 397 column 52 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 iv_b (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 398 column 57 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 399 column 55 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 400 column 58 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out128x6_b (va_get_mem_heaplet 6 va_s0) (va_get_mem_heaplet\n 6 va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 403 column 39 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (plain_num_bytes < pow2_32) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 404 column 38 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRsi va_s0 < pow2_32) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 406 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 408 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (ctr_BE_1:quad32) = iv_BE in label va_range1\n \"***** POSTCONDITION NOT MET AT line 409 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (ctr_BE_2:quad32) = Vale.AES.GCTR_s.inc32 iv_BE 1 in label va_range1\n \"***** POSTCONDITION NOT MET AT line 412 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (plain_in:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b)) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in128_b)) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 421 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (cipher_out:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM) out128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b)) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_sM) out128_b)) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 430 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (cipher_bound:nat) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> len128x6 + len128 + 1) (fun _ -> len128x6 + len128) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 434 column 77 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.AES.GCTR.gctr_partial alg cipher_bound plain_in cipher_out key ctr_BE_2) /\\ label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 438 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (length_quad:quad32) = Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.Def.Types_s.insert_nat64 (Vale.Def.Types_s.insert_nat64 (Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 0 0 0 0) (8 `op_Multiply` va_get_reg64 rRsi va_s0) 1) (8 `op_Multiply`\n plain_num_bytes) 0) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 440 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (raw_auth_quads:(seq quad32)) = va_if (va_get_reg64 rRsi va_s0 > va_get_reg64 rRdx va_s0\n `op_Multiply` 128 `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 7 va_s0) abytes_b)) (fun _ -> Vale.X64.Decls.s128 (va_get_mem_heaplet 1\n va_s0) auth_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 444 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (auth_input_bytes:(seq nat8)) = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_auth_quads) 0 (va_get_reg64 rRsi va_s0) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 445 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (padded_auth_bytes:(seq nat8)) = Vale.AES.GCTR_s.pad_to_128_bits auth_input_bytes in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 446 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (auth_quad_seq:(seq quad32)) = Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_auth_bytes\n in label va_range1\n \"***** POSTCONDITION NOT MET AT line 448 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (raw_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32 (FStar.Seq.Base.append #quad32\n auth_quad_seq (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b))\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 452 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (total_bytes:nat) = FStar.Seq.Base.length #quad32 auth_quad_seq `op_Multiply` 16 +\n plain_num_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 453 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (raw_quad_seq:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply`\n 128 `op_Division` 8) (fun _ -> let (ab:(seq nat8)) = FStar.Seq.Base.slice\n #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes (FStar.Seq.Base.append #quad32\n raw_quad_seq (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b))) 0 total_bytes in let\n (pb:(seq nat8)) = Vale.AES.GCTR_s.pad_to_128_bits ab in Vale.Def.Types_s.le_bytes_to_seq_quad32\n pb) (fun _ -> raw_quad_seq) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 460 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (auth_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32 raw_quad_seq\n (FStar.Seq.Base.create #quad32 1 length_quad) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 461 column 106 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_xmm 8 va_sM == Vale.AES.GCTR_s.gctr_encrypt_block ctr_BE_1 (Vale.AES.GHash_s.ghash_LE\n h_LE auth_quad_seq) alg key 0)))))))))))))))))) in\n assert_norm (va_qc.mods == va_mods);\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet\n 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13;\n va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm\n 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64\n rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64\n rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64\n rRax; va_Mod_ok; va_Mod_mem]) va_sM va_s0;\n (va_sM, va_fM)\n\n[@ va_qattr]\nlet va_wp_Gcm_blocks (alg:algorithm) (offset:int) (auth_b:buffer128) (abytes_b:buffer128)\n (in128x6_b:buffer128) (out128x6_b:buffer128) (in128_b:buffer128) (out128_b:buffer128)\n (inout_b:buffer128) (iv_b:buffer128) (scratch_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (va_s0:va_state) (va_k:(va_state -> unit ->\n Type0)) : Type0 =\n (va_get_ok va_s0 /\\ (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64\n (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0) in let\n (in128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 8) (va_get_stack va_s0) in let (out128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack va_s0) in\n let (len128x6:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 24) (va_get_stack va_s0) in let (in128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack va_s0) in\n let (out128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 40) (va_get_stack va_s0) in let (len128:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack va_s0) in\n let (inout_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 56) (va_get_stack va_s0) in let (plain_num_bytes:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack va_s0) in\n let (h_LE:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.X64.Decls.buffer128_read hkeys_b 2 (va_get_mem_heaplet 0 va_s0)) in sse_enabled /\\\n movbe_enabled /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 0)\n (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64\n (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 24) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 40) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 56) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet\n 1 va_s0) (va_get_reg64 rRdi va_s0) auth_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0)\n Secret /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 7 va_s0) abytes_ptr abytes_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 2\n va_s0) (va_get_reg64 rR8 va_s0) iv_b 1 (va_get_mem_layout va_s0) Public /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 6 va_s0) in128x6_ptr in128x6_b len128x6\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 6\n va_s0) out128x6_ptr out128x6_b len128x6 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0) in128_ptr in128_b len128\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1\n va_s0) out128_ptr out128_b len128 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) inout_ptr inout_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 3\n va_s0) (va_get_reg64 rRbp va_s0) scratch_b 9 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0) hkeys_b 8\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.buffer_disjoints128 iv_b ([keys_b;\n scratch_b; in128x6_b; out128x6_b; hkeys_b; in128_b; out128_b; inout_b]) /\\\n Vale.X64.Decls.buffer_disjoints128 scratch_b ([keys_b; in128x6_b; out128x6_b; in128_b;\n out128_b; inout_b; hkeys_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128x6_b ([keys_b;\n hkeys_b; in128_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128_b ([keys_b; hkeys_b;\n out128x6_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 inout_b ([keys_b; hkeys_b;\n out128x6_b; out128_b]) /\\ (Vale.X64.Decls.buffers_disjoint128 in128x6_b out128x6_b \\/ in128x6_b\n == out128x6_b) /\\ (Vale.X64.Decls.buffers_disjoint128 in128_b out128_b \\/ in128_b == out128_b)\n /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ in128x6_ptr\n + 16 `op_Multiply` len128x6 < pow2_64 /\\ out128x6_ptr + 16 `op_Multiply` len128x6 < pow2_64 /\\\n in128_ptr + 16 `op_Multiply` len128 < pow2_64 /\\ out128_ptr + 16 `op_Multiply` len128 < pow2_64\n /\\ inout_ptr + 16 < pow2_64 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 auth_b ==\n va_get_reg64 rRdx va_s0 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 abytes_b == 1\n /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128x6_b ==\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out128x6_b /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 out128_b /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128\n in128x6_b == len128x6 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b ==\n len128 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 inout_b == 1 /\\\n plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32 /\\ va_get_reg64 rR9 va_s0 + 32 <\n pow2_64 /\\ Vale.X64.Memory.buffer_addr #Vale.X64.Memory.vuint128 keys_b (va_get_mem_heaplet 0\n va_s0) + 128 < pow2_64 /\\ len128x6 `op_Modulus` 6 == 0 /\\ (len128x6 > 0 ==> len128x6 >= 6) /\\\n 12 + len128x6 + 6 < pow2_32 /\\ (va_mul_nat len128x6 (128 `op_Division` 8) + va_mul_nat len128\n (128 `op_Division` 8) <= plain_num_bytes /\\ plain_num_bytes < va_mul_nat len128x6 (128\n `op_Division` 8) + va_mul_nat len128 (128 `op_Division` 8) + 128 `op_Division` 8) /\\\n (va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) <= va_get_reg64 rRsi va_s0 /\\\n va_get_reg64 rRsi va_s0 < va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) + 128\n `op_Division` 8) /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rRcx va_s0)\n (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 (Vale.AES.AES_s.aes_encrypt_LE alg key\n (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0)))) /\\ (forall (va_x_mem:vale_heap)\n (va_x_rax:nat64) (va_x_rbx:nat64) (va_x_rcx:nat64) (va_x_rdx:nat64) (va_x_rdi:nat64)\n (va_x_rsi:nat64) (va_x_rbp:nat64) (va_x_r8:nat64) (va_x_r9:nat64) (va_x_r10:nat64)\n (va_x_r11:nat64) (va_x_r12:nat64) (va_x_r13:nat64) (va_x_r14:nat64) (va_x_r15:nat64)\n (va_x_xmm0:quad32) (va_x_xmm1:quad32) (va_x_xmm2:quad32) (va_x_xmm3:quad32) (va_x_xmm4:quad32)\n (va_x_xmm5:quad32) (va_x_xmm6:quad32) (va_x_xmm7:quad32) (va_x_xmm8:quad32) (va_x_xmm9:quad32)\n (va_x_xmm10:quad32) (va_x_xmm11:quad32) (va_x_xmm12:quad32) (va_x_xmm13:quad32)\n (va_x_xmm14:quad32) (va_x_xmm15:quad32) (va_x_heap1:vale_heap) (va_x_heap2:vale_heap)\n (va_x_heap3:vale_heap) (va_x_heap5:vale_heap) (va_x_heap6:vale_heap)\n (va_x_efl:Vale.X64.Flags.t) . let va_sM = va_upd_flags va_x_efl (va_upd_mem_heaplet 6\n va_x_heap6 (va_upd_mem_heaplet 5 va_x_heap5 (va_upd_mem_heaplet 3 va_x_heap3\n (va_upd_mem_heaplet 2 va_x_heap2 (va_upd_mem_heaplet 1 va_x_heap1 (va_upd_xmm 15 va_x_xmm15\n (va_upd_xmm 14 va_x_xmm14 (va_upd_xmm 13 va_x_xmm13 (va_upd_xmm 12 va_x_xmm12 (va_upd_xmm 11\n va_x_xmm11 (va_upd_xmm 10 va_x_xmm10 (va_upd_xmm 9 va_x_xmm9 (va_upd_xmm 8 va_x_xmm8\n (va_upd_xmm 7 va_x_xmm7 (va_upd_xmm 6 va_x_xmm6 (va_upd_xmm 5 va_x_xmm5 (va_upd_xmm 4 va_x_xmm4\n (va_upd_xmm 3 va_x_xmm3 (va_upd_xmm 2 va_x_xmm2 (va_upd_xmm 1 va_x_xmm1 (va_upd_xmm 0 va_x_xmm0\n (va_upd_reg64 rR15 va_x_r15 (va_upd_reg64 rR14 va_x_r14 (va_upd_reg64 rR13 va_x_r13\n (va_upd_reg64 rR12 va_x_r12 (va_upd_reg64 rR11 va_x_r11 (va_upd_reg64 rR10 va_x_r10\n (va_upd_reg64 rR9 va_x_r9 (va_upd_reg64 rR8 va_x_r8 (va_upd_reg64 rRbp va_x_rbp (va_upd_reg64\n rRsi va_x_rsi (va_upd_reg64 rRdi va_x_rdi (va_upd_reg64 rRdx va_x_rdx (va_upd_reg64 rRcx\n va_x_rcx (va_upd_reg64 rRbx va_x_rbx (va_upd_reg64 rRax va_x_rax (va_upd_mem va_x_mem\n va_s0))))))))))))))))))))))))))))))))))))) in va_get_ok va_sM /\\ (let\n (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in let (h_LE:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.X64.Decls.buffer128_read hkeys_b 2\n (va_get_mem_heaplet 0 va_s0)) in Vale.X64.Decls.modifies_buffer128 out128_b (va_get_mem_heaplet\n 1 va_s0) (va_get_mem_heaplet 1 va_sM) /\\ Vale.X64.Decls.modifies_buffer128 iv_b\n (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2 va_sM) /\\ Vale.X64.Decls.modifies_buffer128\n scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3 va_sM) /\\\n Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM) /\\ Vale.X64.Decls.modifies_buffer128 out128x6_b (va_get_mem_heaplet 6 va_s0)\n (va_get_mem_heaplet 6 va_sM) /\\ plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32\n /\\ (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in let\n (ctr_BE_1:quad32) = iv_BE in let (ctr_BE_2:quad32) = Vale.AES.GCTR_s.inc32 iv_BE 1 in let\n (plain_in:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b)) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in128_b)) in let (cipher_out:(seq quad32)) = va_if\n (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128 `op_Division` 8) (fun _ ->\n FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_sM) out128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM)\n inout_b)) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM)\n out128_b)) in let (cipher_bound:nat) = va_if (plain_num_bytes > (len128x6 + len128)\n `op_Multiply` 128 `op_Division` 8) (fun _ -> len128x6 + len128 + 1) (fun _ -> len128x6 +\n len128) in Vale.AES.GCTR.gctr_partial alg cipher_bound plain_in cipher_out key ctr_BE_2 /\\ (let\n (length_quad:quad32) = Vale.Def.Types_s.reverse_bytes_quad32 (Vale.Def.Types_s.insert_nat64\n (Vale.Def.Types_s.insert_nat64 (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0) (8\n `op_Multiply` va_get_reg64 rRsi va_s0) 1) (8 `op_Multiply` plain_num_bytes) 0) in let\n (raw_auth_quads:(seq quad32)) = va_if (va_get_reg64 rRsi va_s0 > va_get_reg64 rRdx va_s0\n `op_Multiply` 128 `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 7 va_s0) abytes_b)) (fun _ -> Vale.X64.Decls.s128 (va_get_mem_heaplet 1\n va_s0) auth_b) in let (auth_input_bytes:(seq nat8)) = FStar.Seq.Base.slice\n #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_auth_quads) 0 (va_get_reg64\n rRsi va_s0) in let (padded_auth_bytes:(seq nat8)) = Vale.AES.GCTR_s.pad_to_128_bits\n auth_input_bytes in let (auth_quad_seq:(seq quad32)) = Vale.Def.Types_s.le_bytes_to_seq_quad32\n padded_auth_bytes in let (raw_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32\n (FStar.Seq.Base.append #quad32 auth_quad_seq (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0)\n in128x6_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b) in let (total_bytes:nat)\n = FStar.Seq.Base.length #quad32 auth_quad_seq `op_Multiply` 16 + plain_num_bytes in let\n (raw_quad_seq:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> let (ab:(seq nat8)) = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes (FStar.Seq.Base.append #quad32 raw_quad_seq\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b))) 0 total_bytes in let (pb:(seq\n nat8)) = Vale.AES.GCTR_s.pad_to_128_bits ab in Vale.Def.Types_s.le_bytes_to_seq_quad32 pb) (fun\n _ -> raw_quad_seq) in let (auth_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32\n raw_quad_seq (FStar.Seq.Base.create #quad32 1 length_quad) in va_get_xmm 8 va_sM ==\n Vale.AES.GCTR_s.gctr_encrypt_block ctr_BE_1 (Vale.AES.GHash_s.ghash_LE h_LE auth_quad_seq) alg\n key 0))) ==> va_k va_sM (())))\n\n\nval va_wpProof_Gcm_blocks : alg:algorithm -> offset:int -> auth_b:buffer128 -> abytes_b:buffer128\n -> in128x6_b:buffer128 -> out128x6_b:buffer128 -> in128_b:buffer128 -> out128_b:buffer128 ->\n inout_b:buffer128 -> iv_b:buffer128 -> scratch_b:buffer128 -> key:(seq nat32) -> round_keys:(seq\n quad32) -> keys_b:buffer128 -> hkeys_b:buffer128 -> va_s0:va_state -> va_k:(va_state -> unit ->\n Type0)\n -> Ghost (va_state & va_fuel & unit)\n (requires (va_t_require va_s0 /\\ va_wp_Gcm_blocks alg offset auth_b abytes_b in128x6_b out128x6_b\n in128_b out128_b inout_b iv_b scratch_b key round_keys keys_b hkeys_b va_s0 va_k))\n (ensures (fun (va_sM, va_f0, va_g) -> va_t_ensure (va_code_Gcm_blocks alg offset) ([va_Mod_flags;\n va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet 3; va_Mod_mem_heaplet 2;\n va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13; va_Mod_xmm 12; va_Mod_xmm\n 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5;\n va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR15;\n va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64 rR11; va_Mod_reg64 rR10;\n va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64 rRsi; va_Mod_reg64 rRdi;\n va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64 rRax; va_Mod_mem]) va_s0\n va_k ((va_sM, va_f0, va_g))))\n[@\"opaque_to_smt\"]\nlet va_wpProof_Gcm_blocks alg offset auth_b abytes_b in128x6_b out128x6_b in128_b out128_b inout_b\n iv_b scratch_b key round_keys keys_b hkeys_b va_s0 va_k =\n let (va_sM, va_f0) = va_lemma_Gcm_blocks (va_code_Gcm_blocks alg offset) va_s0 alg offset auth_b\n abytes_b in128x6_b out128x6_b in128_b out128_b inout_b iv_b scratch_b key round_keys keys_b\n hkeys_b in\n va_lemma_upd_update va_sM;\n assert (va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 6 va_sM\n (va_update_mem_heaplet 5 va_sM (va_update_mem_heaplet 3 va_sM (va_update_mem_heaplet 2 va_sM\n (va_update_mem_heaplet 1 va_sM (va_update_xmm 15 va_sM (va_update_xmm 14 va_sM (va_update_xmm\n 13 va_sM (va_update_xmm 12 va_sM (va_update_xmm 11 va_sM (va_update_xmm 10 va_sM (va_update_xmm\n 9 va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5\n va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1\n va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR15 va_sM (va_update_reg64 rR14 va_sM\n (va_update_reg64 rR13 va_sM (va_update_reg64 rR12 va_sM (va_update_reg64 rR11 va_sM\n (va_update_reg64 rR10 va_sM (va_update_reg64 rR9 va_sM (va_update_reg64 rR8 va_sM\n (va_update_reg64 rRbp va_sM (va_update_reg64 rRsi va_sM (va_update_reg64 rRdi va_sM\n (va_update_reg64 rRdx va_sM (va_update_reg64 rRcx va_sM (va_update_reg64 rRbx va_sM\n (va_update_reg64 rRax va_sM (va_update_ok va_sM (va_update_mem va_sM\n va_s0))))))))))))))))))))))))))))))))))))))));\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet\n 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13;\n va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm\n 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64\n rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64\n rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64\n rRax; va_Mod_mem]) va_sM va_s0;\n let va_g = () in\n (va_sM, va_f0, va_g)\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_quick_Gcm_blocks (alg:algorithm) (offset:int) (auth_b:buffer128) (abytes_b:buffer128)\n (in128x6_b:buffer128) (out128x6_b:buffer128) (in128_b:buffer128) (out128_b:buffer128)\n (inout_b:buffer128) (iv_b:buffer128) (scratch_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) : (va_quickCode unit (va_code_Gcm_blocks alg\n offset)) =\n (va_QProc (va_code_Gcm_blocks alg offset) ([va_Mod_flags; va_Mod_mem_heaplet 6;\n va_Mod_mem_heaplet 5; va_Mod_mem_heaplet 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1;\n va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13; va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10;\n va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm\n 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64\n rR13; va_Mod_reg64 rR12; va_Mod_reg64 rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64\n rR8; va_Mod_reg64 rRbp; va_Mod_reg64 rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64\n rRcx; va_Mod_reg64 rRbx; va_Mod_reg64 rRax; va_Mod_mem]) (va_wp_Gcm_blocks alg offset auth_b\n abytes_b in128x6_b out128x6_b in128_b out128_b inout_b iv_b scratch_b key round_keys keys_b\n hkeys_b) (va_wpProof_Gcm_blocks alg offset auth_b abytes_b in128x6_b out128x6_b in128_b\n out128_b inout_b iv_b scratch_b key round_keys keys_b hkeys_b))\n#pop-options\n//--\n//-- Gcm_blocks_wrapped\n#push-options \"--z3rlimit 60\"\n\nval va_code_Gcm_blocks_wrapped : alg:algorithm -> offset:int -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_blocks_wrapped alg offset =\n (va_Block (va_CCons (va_code_Gcm_blocks alg offset) (va_CCons (va_Block (va_CNil ())) (va_CCons\n (va_Block (va_CNil ())) (va_CNil ())))))\n\nval va_codegen_success_Gcm_blocks_wrapped : alg:algorithm -> offset:int -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_blocks_wrapped alg offset =\n (va_pbool_and (va_codegen_success_Gcm_blocks alg offset) (va_ttrue ()))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_blocks_wrapped (va_mods:va_mods_t) (alg:algorithm) (offset:int) (auth_b:buffer128)\n (abytes_b:buffer128) (in128x6_b:buffer128) (out128x6_b:buffer128) (in128_b:buffer128)\n (out128_b:buffer128) (inout_b:buffer128) (iv_b:buffer128) (iv:supported_iv_LE)\n (scratch_b:buffer128) (key:(seq nat32)) (round_keys:(seq quad32)) (keys_b:buffer128)\n (hkeys_b:buffer128) (expected_tag:(seq nat8)) : (va_quickCode unit (va_code_Gcm_blocks_wrapped\n alg offset)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in let\n (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 0) (va_get_stack va_s) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 8) (va_get_stack va_s) in let\n (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 16) (va_get_stack va_s) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 24) (va_get_stack va_s) in let\n (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 32) (va_get_stack va_s) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 40) (va_get_stack va_s) in let\n (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset\n + 48) (va_get_stack va_s) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 56) (va_get_stack va_s) in let\n (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s\n + offset + 64) (va_get_stack va_s) in va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 739 column 15 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_blocks alg offset auth_b abytes_b in128x6_b out128x6_b in128_b out128_b inout_b\n iv_b scratch_b key round_keys keys_b hkeys_b) (fun (va_s:va_state) _ -> va_qAssertSquash\n va_range1\n \"***** EXPRESSION PRECONDITIONS NOT MET WITHIN line 741 column 37 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (8 `op_Multiply` va_get_reg64 rRsi va_old_s >= 0 /\\ 8 `op_Multiply` va_get_reg64 rRsi va_old_s\n <= 18446744073709551615 /\\ 8 `op_Multiply` plain_num_bytes >= 0 /\\ 8 `op_Multiply`\n plain_num_bytes <= 18446744073709551615) (fun _ -> let (va_arg55:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.Def.Types_s.insert_nat64\n (Vale.Def.Types_s.insert_nat64 (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0) (8\n `op_Multiply` va_get_reg64 rRsi va_old_s) 1) (8 `op_Multiply` plain_num_bytes) 0) in let\n (va_arg54:Vale.Def.Types_s.quad32) = va_get_xmm 8 va_s in let\n (va_arg53:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.X64.Decls.buffer128_read hkeys_b 2 (va_get_mem_heaplet 0 va_s)) in let\n (va_arg52:Vale.Def.Types_s.quad32) = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2\n va_old_s) in let (va_arg51:Vale.AES.GCM_s.supported_iv_LE) = iv in let (va_arg50:Prims.nat) =\n va_get_reg64 rRsi va_old_s in let (va_arg49:Prims.nat) = plain_num_bytes in let\n (va_arg48:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s) inout_b in let (va_arg47:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) out128_b in let\n (va_arg46:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_s) out128x6_b in let (va_arg45:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_old_s) inout_b in let\n (va_arg44:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_old_s) in128_b in let (va_arg43:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b in\n let (va_arg42:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 7 va_old_s) abytes_b in let (va_arg41:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) auth_b in let\n (va_arg40:(FStar.Seq.Base.seq Vale.Def.Words_s.nat32)) = key in let\n (va_arg39:Vale.AES.AES_common_s.algorithm) = alg in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 741 column 37 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCM.gcm_blocks_dec_helper_simplified va_arg39 va_arg40 va_arg41\n va_arg42 va_arg43 va_arg44 va_arg45 va_arg46 va_arg47 va_arg48 va_arg49 va_arg50 va_arg51\n va_arg52 va_arg53 va_arg54 va_arg55) (let (auth_raw_quads:(FStar.Seq.Base.seq\n Vale.X64.Decls.quad32)) = FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_old_s) auth_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 7 va_old_s)\n abytes_b) in va_qAssertSquash va_range1\n \"***** EXPRESSION PRECONDITIONS NOT MET WITHIN line 751 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n ((fun a_1906 (s_1907:(FStar.Seq.Base.seq a_1906)) (i_1908:Prims.nat) (j_1909:Prims.nat) -> let\n (j_1869:Prims.nat) = j_1909 in Prims.b2t (Prims.op_AmpAmp (Prims.op_LessThanOrEqual i_1908\n j_1869) (Prims.op_LessThanOrEqual j_1869 (FStar.Seq.Base.length #a_1906 s_1907))))\n Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes auth_raw_quads) 0 (va_get_reg64\n rRsi va_old_s)) (fun _ -> let (auth_bytes:(FStar.Seq.Base.seq Vale.Def.Types_s.nat8)) =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n auth_raw_quads) 0 (va_get_reg64 rRsi va_old_s) in let (va_arg38:Vale.Def.Types_s.quad32) =\n Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_old_s) in let\n (va_arg37:Vale.AES.GCM_s.supported_iv_LE) = iv in let (va_arg36:Prims.nat) = plain_num_bytes in\n let (va_arg35:(FStar.Seq.Base.seq Vale.Def.Words_s.nat8)) = expected_tag in let\n (va_arg34:(FStar.Seq.Base.seq Vale.Def.Words_s.nat8)) = auth_bytes in let\n (va_arg33:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s) inout_b in let (va_arg32:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) out128_b in let\n (va_arg31:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_s) out128x6_b in let (va_arg30:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_old_s) inout_b in let\n (va_arg29:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_old_s) in128_b in let (va_arg28:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b in\n let (va_arg27:(FStar.Seq.Base.seq Vale.Def.Words_s.nat32)) = key in let\n (va_arg26:Vale.AES.AES_common_s.algorithm) = alg in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 752 column 37 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCM.gcm_blocks_helper_dec_simplified va_arg26 va_arg27 va_arg28\n va_arg29 va_arg30 va_arg31 va_arg32 va_arg33 va_arg34 va_arg35 va_arg36 va_arg37 va_arg38)\n (va_QEmpty (()))))))))\n\n\nval va_lemma_Gcm_blocks_wrapped : va_b0:va_code -> va_s0:va_state -> alg:algorithm -> offset:int ->\n auth_b:buffer128 -> abytes_b:buffer128 -> in128x6_b:buffer128 -> out128x6_b:buffer128 ->\n in128_b:buffer128 -> out128_b:buffer128 -> inout_b:buffer128 -> iv_b:buffer128 ->\n iv:supported_iv_LE -> scratch_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) ->\n keys_b:buffer128 -> hkeys_b:buffer128 -> expected_tag:(seq nat8)\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_blocks_wrapped alg offset) va_s0 /\\ va_get_ok\n va_s0 /\\ (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64\n rRsp va_s0 + offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in sse_enabled /\\ movbe_enabled /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0)\n Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp\n va_s0 + offset + 8) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 24) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 40) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 56) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet\n 1 va_s0) (va_get_reg64 rRdi va_s0) auth_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0)\n Secret /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 7 va_s0) abytes_ptr abytes_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 2\n va_s0) (va_get_reg64 rR8 va_s0) iv_b 1 (va_get_mem_layout va_s0) Public /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 6 va_s0) in128x6_ptr in128x6_b len128x6\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 6\n va_s0) out128x6_ptr out128x6_b len128x6 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0) in128_ptr in128_b len128\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1\n va_s0) out128_ptr out128_b len128 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) inout_ptr inout_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 3\n va_s0) (va_get_reg64 rRbp va_s0) scratch_b 9 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0) hkeys_b 8\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.buffer_disjoints128 iv_b ([keys_b;\n scratch_b; in128x6_b; out128x6_b; hkeys_b; in128_b; out128_b; inout_b]) /\\\n Vale.X64.Decls.buffer_disjoints128 scratch_b ([keys_b; in128x6_b; out128x6_b; in128_b;\n out128_b; inout_b; hkeys_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128x6_b ([keys_b;\n hkeys_b; in128_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128_b ([keys_b; hkeys_b;\n out128x6_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 inout_b ([keys_b; hkeys_b;\n out128x6_b; out128_b]) /\\ (Vale.X64.Decls.buffers_disjoint128 in128x6_b out128x6_b \\/ in128x6_b\n == out128x6_b) /\\ (Vale.X64.Decls.buffers_disjoint128 in128_b out128_b \\/ in128_b == out128_b)\n /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ in128x6_ptr\n + 16 `op_Multiply` len128x6 < pow2_64 /\\ out128x6_ptr + 16 `op_Multiply` len128x6 < pow2_64 /\\\n in128_ptr + 16 `op_Multiply` len128 < pow2_64 /\\ out128_ptr + 16 `op_Multiply` len128 < pow2_64\n /\\ inout_ptr + 16 < pow2_64 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 auth_b ==\n va_get_reg64 rRdx va_s0 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 abytes_b == 1\n /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128x6_b ==\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out128x6_b /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 out128_b /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128\n in128x6_b == len128x6 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b ==\n len128 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 inout_b == 1 /\\\n plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32 /\\ va_get_reg64 rR9 va_s0 + 32 <\n pow2_64 /\\ Vale.X64.Memory.buffer_addr #Vale.X64.Memory.vuint128 keys_b (va_get_mem_heaplet 0\n va_s0) + 128 < pow2_64 /\\ len128x6 `op_Modulus` 6 == 0 /\\ (len128x6 > 0 ==> len128x6 >= 6) /\\\n 12 + len128x6 + 6 < pow2_32 /\\ (va_mul_nat len128x6 (128 `op_Division` 8) + va_mul_nat len128\n (128 `op_Division` 8) <= plain_num_bytes /\\ plain_num_bytes < va_mul_nat len128x6 (128\n `op_Division` 8) + va_mul_nat len128 (128 `op_Division` 8) + 128 `op_Division` 8) /\\\n (va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) <= va_get_reg64 rRsi va_s0 /\\\n va_get_reg64 rRsi va_s0 < va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) + 128\n `op_Division` 8) /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rRcx va_s0)\n (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 (Vale.AES.AES_s.aes_encrypt_LE alg key\n (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0))) /\\ (let iv_BE =\n Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in let h_LE =\n Vale.AES.AES_s.aes_encrypt_LE alg key (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0)\n in iv_BE == Vale.AES.GCM_s.compute_iv_BE h_LE iv))))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in Vale.X64.Decls.modifies_buffer128 out128_b\n (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1 va_sM) /\\ Vale.X64.Decls.modifies_buffer128\n iv_b (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2 va_sM) /\\\n Vale.X64.Decls.modifies_buffer128 scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3\n va_sM) /\\ Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0)\n (va_get_mem_heaplet 5 va_sM) /\\ Vale.X64.Decls.modifies_buffer128 out128x6_b\n (va_get_mem_heaplet 6 va_s0) (va_get_mem_heaplet 6 va_sM) /\\ plain_num_bytes < pow2_32 /\\\n va_get_reg64 rRsi va_s0 < pow2_32 /\\ (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0\n (va_get_mem_heaplet 2 va_s0) in let auth_raw_quads = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 7 va_s0) abytes_b) in let auth_bytes =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n auth_raw_quads) 0 (va_get_reg64 rRsi va_s0) in let plain_raw_quads = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_s0) in128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0)\n in128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b) in let plain_bytes =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n plain_raw_quads) 0 plain_num_bytes in let cipher_raw_quads = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM)\n out128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM) inout_b) in let cipher_bytes =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n cipher_raw_quads) 0 plain_num_bytes in l_and (l_and (l_and (FStar.Seq.Base.length\n #Vale.Def.Types_s.nat8 auth_bytes < pow2_32) (FStar.Seq.Base.length #Vale.Def.Types_s.nat8\n plain_bytes < pow2_32)) (cipher_bytes == __proj__Mktuple2__item___1 #(FStar.Seq.Base.seq\n Vale.Def.Types_s.nat8) #bool (Vale.AES.GCM_s.gcm_decrypt_LE alg\n (Vale.Def.Words.Seq_s.seq_nat32_to_seq_nat8_LE key) iv plain_bytes auth_bytes expected_tag)))\n (Vale.Def.Types_s.le_quad32_to_bytes (va_get_xmm 8 va_sM) == Vale.AES.GCM.gcm_decrypt_LE_tag\n alg (Vale.Def.Words.Seq_s.seq_nat32_to_seq_nat8_LE key) iv plain_bytes auth_bytes))) /\\\n va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 6 va_sM (va_update_mem_heaplet\n 5 va_sM (va_update_mem_heaplet 3 va_sM (va_update_mem_heaplet 2 va_sM (va_update_mem_heaplet 1\n va_sM (va_update_xmm 15 va_sM (va_update_xmm 14 va_sM (va_update_xmm 13 va_sM (va_update_xmm 12\n va_sM (va_update_xmm 11 va_sM (va_update_xmm 10 va_sM (va_update_xmm 9 va_sM (va_update_xmm 8\n va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5 va_sM (va_update_xmm 4\n va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1 va_sM (va_update_xmm 0\n va_sM (va_update_reg64 rR15 va_sM (va_update_reg64 rR14 va_sM (va_update_reg64 rR13 va_sM\n (va_update_reg64 rR12 va_sM (va_update_reg64 rR11 va_sM (va_update_reg64 rR10 va_sM\n (va_update_reg64 rR9 va_sM (va_update_reg64 rR8 va_sM (va_update_reg64 rRbp va_sM\n (va_update_reg64 rRsi va_sM (va_update_reg64 rRdi va_sM (va_update_reg64 rRdx va_sM\n (va_update_reg64 rRcx va_sM (va_update_reg64 rRbx va_sM (va_update_reg64 rRax va_sM\n (va_update_ok va_sM (va_update_mem va_sM va_s0)))))))))))))))))))))))))))))))))))))))))\n[@\"opaque_to_smt\"]\nlet va_lemma_Gcm_blocks_wrapped va_b0 va_s0 alg offset auth_b abytes_b in128x6_b out128x6_b in128_b\n out128_b inout_b iv_b iv scratch_b key round_keys keys_b hkeys_b expected_tag =\n let (va_mods:va_mods_t) = [va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5;\n va_Mod_mem_heaplet 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14;\n va_Mod_xmm 13; va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8;\n va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm\n 1; va_Mod_xmm 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12;\n va_Mod_reg64 rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp;\n va_Mod_reg64 rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx;\n va_Mod_reg64 rRax; va_Mod_ok; va_Mod_mem] in\n let va_qc = va_qcode_Gcm_blocks_wrapped va_mods alg offset auth_b abytes_b in128x6_b out128x6_b\n in128_b out128_b inout_b iv_b iv scratch_b key round_keys keys_b hkeys_b expected_tag in\n let (va_sM, va_fM, va_g) = va_wp_sound_code_norm (va_code_Gcm_blocks_wrapped alg offset) va_qc\n va_s0 (fun va_s0 va_sM va_g -> let () = va_g in label va_range1\n \"***** POSTCONDITION NOT MET AT line 588 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\ (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64\n (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0) in let\n (in128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 8) (va_get_stack va_s0) in let (out128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack va_s0) in\n let (len128x6:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 24) (va_get_stack va_s0) in let (in128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack va_s0) in\n let (out128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 40) (va_get_stack va_s0) in let (len128:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack va_s0) in\n let (inout_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 56) (va_get_stack va_s0) in let (plain_num_bytes:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack va_s0) in\n label va_range1\n \"***** POSTCONDITION NOT MET AT line 711 column 56 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out128_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 712 column 52 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 iv_b (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 713 column 57 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 714 column 55 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 715 column 58 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out128x6_b (va_get_mem_heaplet 6 va_s0) (va_get_mem_heaplet\n 6 va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 718 column 39 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (plain_num_bytes < pow2_32) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 719 column 38 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRsi va_s0 < pow2_32) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 721 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 723 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let auth_raw_quads = FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) auth_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 7 va_s0)\n abytes_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 724 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let auth_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes auth_raw_quads) 0 (va_get_reg64 rRsi va_s0) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 725 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let plain_raw_quads = FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 726 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let plain_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes plain_raw_quads) 0 plain_num_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 727 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let cipher_raw_quads = FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM) out128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 728 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let cipher_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes cipher_raw_quads) 0 plain_num_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 737 column 52 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (l_and (l_and (l_and (FStar.Seq.Base.length #Vale.Def.Types_s.nat8 auth_bytes < pow2_32)\n (FStar.Seq.Base.length #Vale.Def.Types_s.nat8 plain_bytes < pow2_32)) (cipher_bytes ==\n __proj__Mktuple2__item___1 #(FStar.Seq.Base.seq Vale.Def.Types_s.nat8) #bool\n (Vale.AES.GCM_s.gcm_decrypt_LE alg (Vale.Def.Words.Seq_s.seq_nat32_to_seq_nat8_LE key) iv\n plain_bytes auth_bytes expected_tag))) (Vale.Def.Types_s.le_quad32_to_bytes (va_get_xmm 8\n va_sM) == Vale.AES.GCM.gcm_decrypt_LE_tag alg (Vale.Def.Words.Seq_s.seq_nat32_to_seq_nat8_LE\n key) iv plain_bytes auth_bytes))))))))))) in\n assert_norm (va_qc.mods == va_mods);\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet\n 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13;\n va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm\n 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64\n rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64\n rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64\n rRax; va_Mod_ok; va_Mod_mem]) va_sM va_s0;\n (va_sM, va_fM)\n\n[@ va_qattr]\nlet va_wp_Gcm_blocks_wrapped (alg:algorithm) (offset:int) (auth_b:buffer128) (abytes_b:buffer128)\n (in128x6_b:buffer128) (out128x6_b:buffer128) (in128_b:buffer128) (out128_b:buffer128)\n (inout_b:buffer128) (iv_b:buffer128) (iv:supported_iv_LE) (scratch_b:buffer128) (key:(seq nat32))\n (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (expected_tag:(seq nat8))\n (va_s0:va_state) (va_k:(va_state -> unit -> Type0)) : Type0 =\n (va_get_ok va_s0 /\\ (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64\n (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0) in let\n (in128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 8) (va_get_stack va_s0) in let (out128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack va_s0) in\n let (len128x6:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 24) (va_get_stack va_s0) in let (in128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack va_s0) in\n let (out128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 40) (va_get_stack va_s0) in let (len128:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack va_s0) in\n let (inout_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 56) (va_get_stack va_s0) in let (plain_num_bytes:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack va_s0) in\n sse_enabled /\\ movbe_enabled /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 +\n offset + 0) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0)\n Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 32) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 48) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 64) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0) (va_get_reg64 rRdi va_s0) auth_b\n (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 7 va_s0) abytes_ptr abytes_b 1 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 2 va_s0) (va_get_reg64 rR8 va_s0) iv_b 1\n (va_get_mem_layout va_s0) Public /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 6\n va_s0) in128x6_ptr in128x6_b len128x6 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 6 va_s0) out128x6_ptr out128x6_b len128x6\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1\n va_s0) in128_ptr in128_b len128 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1 va_s0) out128_ptr out128_b len128\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5\n va_s0) inout_ptr inout_b 1 (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128\n (va_get_mem_heaplet 3 va_s0) (va_get_reg64 rRbp va_s0) scratch_b 9 (va_get_mem_layout va_s0)\n Secret /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0)\n hkeys_b 8 (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.buffer_disjoints128 iv_b ([keys_b;\n scratch_b; in128x6_b; out128x6_b; hkeys_b; in128_b; out128_b; inout_b]) /\\\n Vale.X64.Decls.buffer_disjoints128 scratch_b ([keys_b; in128x6_b; out128x6_b; in128_b;\n out128_b; inout_b; hkeys_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128x6_b ([keys_b;\n hkeys_b; in128_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128_b ([keys_b; hkeys_b;\n out128x6_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 inout_b ([keys_b; hkeys_b;\n out128x6_b; out128_b]) /\\ (Vale.X64.Decls.buffers_disjoint128 in128x6_b out128x6_b \\/ in128x6_b\n == out128x6_b) /\\ (Vale.X64.Decls.buffers_disjoint128 in128_b out128_b \\/ in128_b == out128_b)\n /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ in128x6_ptr\n + 16 `op_Multiply` len128x6 < pow2_64 /\\ out128x6_ptr + 16 `op_Multiply` len128x6 < pow2_64 /\\\n in128_ptr + 16 `op_Multiply` len128 < pow2_64 /\\ out128_ptr + 16 `op_Multiply` len128 < pow2_64\n /\\ inout_ptr + 16 < pow2_64 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 auth_b ==\n va_get_reg64 rRdx va_s0 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 abytes_b == 1\n /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128x6_b ==\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out128x6_b /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 out128_b /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128\n in128x6_b == len128x6 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b ==\n len128 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 inout_b == 1 /\\\n plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32 /\\ va_get_reg64 rR9 va_s0 + 32 <\n pow2_64 /\\ Vale.X64.Memory.buffer_addr #Vale.X64.Memory.vuint128 keys_b (va_get_mem_heaplet 0\n va_s0) + 128 < pow2_64 /\\ len128x6 `op_Modulus` 6 == 0 /\\ (len128x6 > 0 ==> len128x6 >= 6) /\\\n 12 + len128x6 + 6 < pow2_32 /\\ (va_mul_nat len128x6 (128 `op_Division` 8) + va_mul_nat len128\n (128 `op_Division` 8) <= plain_num_bytes /\\ plain_num_bytes < va_mul_nat len128x6 (128\n `op_Division` 8) + va_mul_nat len128 (128 `op_Division` 8) + 128 `op_Division` 8) /\\\n (va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) <= va_get_reg64 rRsi va_s0 /\\\n va_get_reg64 rRsi va_s0 < va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) + 128\n `op_Division` 8) /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rRcx va_s0)\n (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 (Vale.AES.AES_s.aes_encrypt_LE alg key\n (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0))) /\\ (let iv_BE =\n Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in let h_LE =\n Vale.AES.AES_s.aes_encrypt_LE alg key (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0)\n in iv_BE == Vale.AES.GCM_s.compute_iv_BE h_LE iv)) /\\ (forall (va_x_mem:vale_heap)\n (va_x_rax:nat64) (va_x_rbx:nat64) (va_x_rcx:nat64) (va_x_rdx:nat64) (va_x_rdi:nat64)\n (va_x_rsi:nat64) (va_x_rbp:nat64) (va_x_r8:nat64) (va_x_r9:nat64) (va_x_r10:nat64)\n (va_x_r11:nat64) (va_x_r12:nat64) (va_x_r13:nat64) (va_x_r14:nat64) (va_x_r15:nat64)\n (va_x_xmm0:quad32) (va_x_xmm1:quad32) (va_x_xmm2:quad32) (va_x_xmm3:quad32) (va_x_xmm4:quad32)\n (va_x_xmm5:quad32) (va_x_xmm6:quad32) (va_x_xmm7:quad32) (va_x_xmm8:quad32) (va_x_xmm9:quad32)\n (va_x_xmm10:quad32) (va_x_xmm11:quad32) (va_x_xmm12:quad32) (va_x_xmm13:quad32)\n (va_x_xmm14:quad32) (va_x_xmm15:quad32) (va_x_heap1:vale_heap) (va_x_heap2:vale_heap)\n (va_x_heap3:vale_heap) (va_x_heap5:vale_heap) (va_x_heap6:vale_heap)\n (va_x_efl:Vale.X64.Flags.t) . let va_sM = va_upd_flags va_x_efl (va_upd_mem_heaplet 6\n va_x_heap6 (va_upd_mem_heaplet 5 va_x_heap5 (va_upd_mem_heaplet 3 va_x_heap3\n (va_upd_mem_heaplet 2 va_x_heap2 (va_upd_mem_heaplet 1 va_x_heap1 (va_upd_xmm 15 va_x_xmm15\n (va_upd_xmm 14 va_x_xmm14 (va_upd_xmm 13 va_x_xmm13 (va_upd_xmm 12 va_x_xmm12 (va_upd_xmm 11\n va_x_xmm11 (va_upd_xmm 10 va_x_xmm10 (va_upd_xmm 9 va_x_xmm9 (va_upd_xmm 8 va_x_xmm8\n (va_upd_xmm 7 va_x_xmm7 (va_upd_xmm 6 va_x_xmm6 (va_upd_xmm 5 va_x_xmm5 (va_upd_xmm 4 va_x_xmm4\n (va_upd_xmm 3 va_x_xmm3 (va_upd_xmm 2 va_x_xmm2 (va_upd_xmm 1 va_x_xmm1 (va_upd_xmm 0 va_x_xmm0\n (va_upd_reg64 rR15 va_x_r15 (va_upd_reg64 rR14 va_x_r14 (va_upd_reg64 rR13 va_x_r13\n (va_upd_reg64 rR12 va_x_r12 (va_upd_reg64 rR11 va_x_r11 (va_upd_reg64 rR10 va_x_r10\n (va_upd_reg64 rR9 va_x_r9 (va_upd_reg64 rR8 va_x_r8 (va_upd_reg64 rRbp va_x_rbp (va_upd_reg64\n rRsi va_x_rsi (va_upd_reg64 rRdi va_x_rdi (va_upd_reg64 rRdx va_x_rdx (va_upd_reg64 rRcx\n va_x_rcx (va_upd_reg64 rRbx va_x_rbx (va_upd_reg64 rRax va_x_rax (va_upd_mem va_x_mem\n va_s0))))))))))))))))))))))))))))))))))))) in va_get_ok va_sM /\\ (let\n (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in Vale.X64.Decls.modifies_buffer128 out128_b\n (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1 va_sM) /\\ Vale.X64.Decls.modifies_buffer128\n iv_b (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2 va_sM) /\\\n Vale.X64.Decls.modifies_buffer128 scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3\n va_sM) /\\ Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0)\n (va_get_mem_heaplet 5 va_sM) /\\ Vale.X64.Decls.modifies_buffer128 out128x6_b\n (va_get_mem_heaplet 6 va_s0) (va_get_mem_heaplet 6 va_sM) /\\ plain_num_bytes < pow2_32 /\\\n va_get_reg64 rRsi va_s0 < pow2_32 /\\ (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0\n (va_get_mem_heaplet 2 va_s0) in let auth_raw_quads = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 7 va_s0) abytes_b) in let auth_bytes =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n auth_raw_quads) 0 (va_get_reg64 rRsi va_s0) in let plain_raw_quads = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_s0) in128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0)\n in128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b) in let plain_bytes =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n plain_raw_quads) 0 plain_num_bytes in let cipher_raw_quads = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM)\n out128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM) inout_b) in let cipher_bytes =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n cipher_raw_quads) 0 plain_num_bytes in l_and (l_and (l_and (FStar.Seq.Base.length\n #Vale.Def.Types_s.nat8 auth_bytes < pow2_32) (FStar.Seq.Base.length #Vale.Def.Types_s.nat8\n plain_bytes < pow2_32)) (cipher_bytes == __proj__Mktuple2__item___1 #(FStar.Seq.Base.seq\n Vale.Def.Types_s.nat8) #bool (Vale.AES.GCM_s.gcm_decrypt_LE alg\n (Vale.Def.Words.Seq_s.seq_nat32_to_seq_nat8_LE key) iv plain_bytes auth_bytes expected_tag)))\n (Vale.Def.Types_s.le_quad32_to_bytes (va_get_xmm 8 va_sM) == Vale.AES.GCM.gcm_decrypt_LE_tag\n alg (Vale.Def.Words.Seq_s.seq_nat32_to_seq_nat8_LE key) iv plain_bytes auth_bytes))) ==> va_k\n va_sM (())))\n\n\nval va_wpProof_Gcm_blocks_wrapped : alg:algorithm -> offset:int -> auth_b:buffer128 ->\n abytes_b:buffer128 -> in128x6_b:buffer128 -> out128x6_b:buffer128 -> in128_b:buffer128 ->\n out128_b:buffer128 -> inout_b:buffer128 -> iv_b:buffer128 -> iv:supported_iv_LE ->\n scratch_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128 -> expected_tag:(seq nat8) -> va_s0:va_state -> va_k:(va_state -> unit -> Type0)\n -> Ghost (va_state & va_fuel & unit)\n (requires (va_t_require va_s0 /\\ va_wp_Gcm_blocks_wrapped alg offset auth_b abytes_b in128x6_b\n out128x6_b in128_b out128_b inout_b iv_b iv scratch_b key round_keys keys_b hkeys_b\n expected_tag va_s0 va_k))\n (ensures (fun (va_sM, va_f0, va_g) -> va_t_ensure (va_code_Gcm_blocks_wrapped alg offset)\n ([va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet 3;\n va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13;\n va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm\n 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64\n rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64\n rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64\n rRax; va_Mod_mem]) va_s0 va_k ((va_sM, va_f0, va_g))))\n[@\"opaque_to_smt\"]\nlet va_wpProof_Gcm_blocks_wrapped alg offset auth_b abytes_b in128x6_b out128x6_b in128_b out128_b\n inout_b iv_b iv scratch_b key round_keys keys_b hkeys_b expected_tag va_s0 va_k =\n let (va_sM, va_f0) = va_lemma_Gcm_blocks_wrapped (va_code_Gcm_blocks_wrapped alg offset) va_s0\n alg offset auth_b abytes_b in128x6_b out128x6_b in128_b out128_b inout_b iv_b iv scratch_b key\n round_keys keys_b hkeys_b expected_tag in\n va_lemma_upd_update va_sM;\n assert (va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 6 va_sM\n (va_update_mem_heaplet 5 va_sM (va_update_mem_heaplet 3 va_sM (va_update_mem_heaplet 2 va_sM\n (va_update_mem_heaplet 1 va_sM (va_update_xmm 15 va_sM (va_update_xmm 14 va_sM (va_update_xmm\n 13 va_sM (va_update_xmm 12 va_sM (va_update_xmm 11 va_sM (va_update_xmm 10 va_sM (va_update_xmm\n 9 va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5\n va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1\n va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR15 va_sM (va_update_reg64 rR14 va_sM\n (va_update_reg64 rR13 va_sM (va_update_reg64 rR12 va_sM (va_update_reg64 rR11 va_sM\n (va_update_reg64 rR10 va_sM (va_update_reg64 rR9 va_sM (va_update_reg64 rR8 va_sM\n (va_update_reg64 rRbp va_sM (va_update_reg64 rRsi va_sM (va_update_reg64 rRdi va_sM\n (va_update_reg64 rRdx va_sM (va_update_reg64 rRcx va_sM (va_update_reg64 rRbx va_sM\n (va_update_reg64 rRax va_sM (va_update_ok va_sM (va_update_mem va_sM\n va_s0))))))))))))))))))))))))))))))))))))))));\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet\n 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13;\n va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm\n 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64\n rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64\n rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64\n rRax; va_Mod_mem]) va_sM va_s0;\n let va_g = () in\n (va_sM, va_f0, va_g)\n\nPartial implementation:\nlet va_quick_Gcm_blocks_wrapped\n (alg: algorithm)\n (offset: int)\n (auth_b abytes_b in128x6_b out128x6_b in128_b out128_b inout_b iv_b: buffer128)\n (iv: supported_iv_LE)\n (scratch_b: buffer128)\n (key: (seq nat32))\n (round_keys: (seq quad32))\n (keys_b hkeys_b: buffer128)\n (expected_tag: (seq nat8))\n : (va_quickCode unit (va_code_Gcm_blocks_wrapped alg offset)) =\n \n"
},
{
"role": "assistant",
"content": "(va_QProc (va_code_Gcm_blocks_wrapped alg offset)\n ([\n va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet 3;\n va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13;\n va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1;\n va_Mod_xmm 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12;\n va_Mod_reg64 rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp;\n va_Mod_reg64 rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx;\n va_Mod_reg64 rRbx; va_Mod_reg64 rRax; va_Mod_mem\n ])\n (va_wp_Gcm_blocks_wrapped alg offset auth_b abytes_b in128x6_b out128x6_b in128_b out128_b\n inout_b iv_b iv scratch_b key round_keys keys_b hkeys_b expected_tag)\n (va_wpProof_Gcm_blocks_wrapped alg offset auth_b abytes_b in128x6_b out128x6_b in128_b out128_b\n inout_b iv_b iv scratch_b key round_keys keys_b hkeys_b expected_tag))"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n va_b0: Vale.X64.Decls.va_code ->\n va_s0: Vale.X64.Decls.va_state ->\n alg: Vale.AES.AES_common_s.algorithm ->\n inout_b: Vale.X64.Memory.buffer128 ->\n key: FStar.Seq.Base.seq Vale.X64.Memory.nat32 ->\n round_keys: FStar.Seq.Base.seq Vale.X64.Decls.quad32 ->\n keys_b: Vale.X64.Memory.buffer128 ->\n hkeys_b: Vale.X64.Memory.buffer128 ->\n total_bytes: Prims.nat ->\n old_hash: Vale.X64.Decls.quad32 ->\n completed_quads: FStar.Seq.Base.seq Vale.X64.Decls.quad32 ->\n h_LE: Vale.X64.Decls.quad32\n -> Prims.Ghost (Vale.X64.Decls.va_state * Vale.X64.Decls.va_fuel)\nEffect: Prims.Ghost\n\nContext:\nmodule Vale.AES.X64.GCMdecryptOpt\nopen Vale.Def.Prop_s\nopen Vale.Def.Opaque_s\nopen FStar.Seq\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Seq_s\nopen Vale.Def.Types_s\nopen Vale.Arch.Types\nopen Vale.Arch.HeapImpl\nopen Vale.AES.AES_s\nopen Vale.AES.GCTR_s\nopen Vale.AES.GCTR\nopen Vale.AES.GCM\nopen Vale.AES.GHash_s\nopen Vale.AES.GHash\nopen Vale.AES.GCM_s\nopen Vale.AES.X64.AES\nopen Vale.AES.GF128_s\nopen Vale.AES.GF128\nopen Vale.Poly1305.Math\nopen Vale.AES.GCM_helpers\nopen Vale.AES.X64.GHash\nopen Vale.AES.X64.GCTR\nopen Vale.X64.Machine_s\nopen Vale.X64.Memory\nopen Vale.X64.Stack_i\nopen Vale.X64.State\nopen Vale.X64.Decls\nopen Vale.X64.InsBasic\nopen Vale.X64.InsMem\nopen Vale.X64.InsVector\nopen Vale.X64.InsStack\nopen Vale.X64.InsAes\nopen Vale.X64.QuickCode\nopen Vale.X64.QuickCodes\nopen Vale.AES.X64.GF128_Mul\nopen Vale.X64.Stack\nopen Vale.X64.CPU_Features_s\nopen Vale.Math.Poly2.Bits_s\nopen Vale.AES.X64.AESopt\nopen Vale.AES.X64.AESGCM\nopen Vale.AES.X64.AESopt2\nopen Vale.Lib.Meta\nopen Vale.AES.X64.GCMencryptOpt\nopen Vale.AES.OptPublic\nopen Vale.Lib.Basic\n#reset-options \"--z3rlimit 20 --max_ifuel 0\"\n//-- Gcm_extra_bytes\n\nval va_code_Gcm_extra_bytes : alg:algorithm -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_extra_bytes alg =\n (va_Block (va_CCons (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0)\n (va_op_reg_opr64_reg64 rRax) 0 Secret) (va_CCons (va_code_Mov128 (va_op_xmm_xmm 10)\n (va_op_xmm_xmm 0)) (va_CCons (va_code_Ghash_extra_bytes ()) (va_CCons (va_code_Mov128\n (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_CCons (va_code_Pshufb (va_op_xmm_xmm 0)\n (va_op_xmm_xmm 9)) (va_CCons (va_code_AESEncryptBlock alg) (va_CCons (va_code_Pxor\n (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (va_CCons (va_code_Store128_buffer\n (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax) (va_op_xmm_xmm 10) 0 Secret)\n (va_CNil ()))))))))))\n\nval va_codegen_success_Gcm_extra_bytes : alg:algorithm -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_extra_bytes alg =\n (va_pbool_and (va_codegen_success_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0)\n (va_op_reg_opr64_reg64 rRax) 0 Secret) (va_pbool_and (va_codegen_success_Mov128 (va_op_xmm_xmm\n 10) (va_op_xmm_xmm 0)) (va_pbool_and (va_codegen_success_Ghash_extra_bytes ()) (va_pbool_and\n (va_codegen_success_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_pbool_and\n (va_codegen_success_Pshufb (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (va_pbool_and\n (va_codegen_success_AESEncryptBlock alg) (va_pbool_and (va_codegen_success_Pxor (va_op_xmm_xmm\n 10) (va_op_xmm_xmm 0)) (va_pbool_and (va_codegen_success_Store128_buffer\n (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax) (va_op_xmm_xmm 10) 0 Secret)\n (va_ttrue ())))))))))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_extra_bytes (va_mods:va_mods_t) (alg:algorithm) (inout_b:buffer128) (key:(seq\n nat32)) (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (total_bytes:nat)\n (old_hash:quad32) (completed_quads:(seq quad32)) (h_LE:quad32) : (va_quickCode unit\n (va_code_Gcm_extra_bytes alg)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in let (len:(va_int_range\n 1 1)) = 1 in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 188 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64\n rRax) 0 Secret inout_b 0) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 189 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov128 (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (fun (va_s:va_state) _ -> let\n (hash_input:quad32) = va_get_xmm 0 va_s in va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 193 column 22 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Ghash_extra_bytes hkeys_b total_bytes old_hash h_LE completed_quads) (fun\n (va_s:va_state) _ -> va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 194 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (FStar.Seq.Base.equal #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s)\n inout_b) (FStar.Seq.Base.create #quad32 1 hash_input)) (let (snap:(FStar.Seq.Base.seq\n Vale.X64.Decls.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s) inout_b in va_QSeq\n va_range1\n \"***** PRECONDITION NOT MET AT line 198 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 199 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Pshufb (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (fun (va_s:va_state) _ -> va_QBind\n va_range1\n \"***** PRECONDITION NOT MET AT line 200 column 20 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_AESEncryptBlock alg (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 11 va_s)) key\n round_keys keys_b) (fun (va_s:va_state) _ -> va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 201 column 26 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.AES_s.aes_encrypt_LE_reveal ()) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 204 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Pxor (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 205 column 20 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Store128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax)\n (va_op_xmm_xmm 10) 0 Secret inout_b 0) (fun (va_s:va_state) _ -> va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 207 column 24 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.gctr_partial_reveal ()) (va_QEmpty (()))))))))))))))\n\n\nval va_lemma_Gcm_extra_bytes : va_b0:va_code -> va_s0:va_state -> alg:algorithm ->\n inout_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128 -> total_bytes:nat -> old_hash:quad32 -> completed_quads:(seq quad32) ->\n h_LE:quad32\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_extra_bytes alg) va_s0 /\\ va_get_ok va_s0 /\\ (let\n (len:(va_int_range 1 1)) = 1 in sse_enabled /\\ Vale.X64.Decls.buffers_disjoint128 keys_b\n inout_b /\\ Vale.X64.Decls.buffers_disjoint128 hkeys_b inout_b /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) (va_get_reg64 rRax va_s0) inout_b\n len (va_get_mem_layout va_s0) Secret /\\ len == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 inout_b /\\ va_get_xmm 9 va_s0 == Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\ aes_reqs alg key round_keys\n keys_b (va_get_reg64 rR8 va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\\n pclmulqdq_enabled /\\ Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0\n va_s0) hkeys_b) (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret /\\ va_get_xmm 8 va_s0 == Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.AES.GHash.ghash_incremental0 h_LE old_hash completed_quads) /\\ FStar.Seq.Base.length\n #quad32 completed_quads == total_bytes `op_Division` 16 /\\ total_bytes < 16 `op_Multiply`\n FStar.Seq.Base.length #quad32 completed_quads + 16 /\\ va_get_reg64 rR10 va_s0 == total_bytes\n `op_Modulus` 16 /\\ total_bytes `op_Modulus` 16 =!= 0 /\\ (0 < total_bytes /\\ total_bytes < 16\n `op_Multiply` Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes) /\\ 16 `op_Multiply`\n (Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes - 1) < total_bytes)))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (let (len:(va_int_range 1 1)) = 1 in Vale.X64.Decls.modifies_buffer128 inout_b\n (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5 va_sM) /\\ Vale.AES.GCTR.gctr_partial alg len\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b) key (va_get_xmm 11 va_s0) /\\ (let raw_quads =\n FStar.Seq.Base.append #quad32 completed_quads (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0)\n inout_b) in let input_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads) 0 total_bytes in let padded_bytes =\n Vale.AES.GCTR_s.pad_to_128_bits input_bytes in let input_quads =\n Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in l_and (FStar.Seq.Base.length\n #Vale.Def.Types_s.quad32 input_quads > 0) (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8\n va_sM) == Vale.AES.GHash.ghash_incremental h_LE old_hash input_quads))) /\\ va_state_eq va_sM\n (va_update_flags va_sM (va_update_mem_heaplet 5 va_sM (va_update_xmm 10 va_sM (va_update_xmm 8\n va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5 va_sM (va_update_xmm 4\n va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1 va_sM (va_update_xmm 0\n va_sM (va_update_reg64 rR11 va_sM (va_update_reg64 rRcx va_sM (va_update_ok va_sM\n (va_update_mem va_sM va_s0))))))))))))))))))\n\nPartial implementation:\nlet va_lemma_Gcm_extra_bytes\n va_b0\n va_s0\n alg\n inout_b\n key\n round_keys\n keys_b\n hkeys_b\n total_bytes\n old_hash\n completed_quads\n h_LE\n =\n \n"
},
{
"role": "assistant",
"content": "let va_mods:va_mods_t =\n [\n va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6;\n va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0;\n va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_ok; va_Mod_mem\n ]\nin\nlet va_qc =\n va_qcode_Gcm_extra_bytes va_mods alg inout_b key round_keys keys_b hkeys_b total_bytes old_hash\n completed_quads h_LE\nin\nlet va_sM, va_fM, va_g =\n va_wp_sound_code_norm (va_code_Gcm_extra_bytes alg)\n va_qc\n va_s0\n (fun va_s0 va_sM va_g ->\n let () = va_g in\n label va_range1\n \"***** POSTCONDITION NOT MET AT line 121 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\\n (let len:(va_int_range 1 1) = 1 in\n label va_range1\n \"***** POSTCONDITION NOT MET AT line 174 column 55 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 inout_b\n (va_get_mem_heaplet 5 va_s0)\n (va_get_mem_heaplet 5 va_sM)) /\\\n label va_range1\n \"***** POSTCONDITION NOT MET AT line 177 column 95 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.AES.GCTR.gctr_partial alg\n len\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM) inout_b)\n key\n (va_get_xmm 11 va_s0)) /\\\n label va_range1\n \"***** POSTCONDITION NOT MET AT line 180 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let raw_quads =\n FStar.Seq.Base.append #quad32\n completed_quads\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b)\n in\n label va_range1\n \"***** POSTCONDITION NOT MET AT line 181 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let input_bytes =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads)\n 0\n total_bytes\n in\n label va_range1\n \"***** POSTCONDITION NOT MET AT line 182 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let padded_bytes = Vale.AES.GCTR_s.pad_to_128_bits input_bytes in\n label va_range1\n \"***** POSTCONDITION NOT MET AT line 183 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let input_quads = Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in\n label va_range1\n \"***** POSTCONDITION NOT MET AT line 186 column 59 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (l_and (FStar.Seq.Base.length #Vale.Def.Types_s.quad32 input_quads > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) ==\n Vale.AES.GHash.ghash_incremental h_LE old_hash input_quads))))))))\nin\nassert_norm (va_qc.mods == va_mods);\nva_lemma_norm_mods ([\n va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6;\n va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0;\n va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_ok; va_Mod_mem\n ])\n va_sM\n va_s0;\n(va_sM, va_fM)"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: pk: Spec.Box.publickey -> sk: Spec.Box.secretkey\n -> FStar.Pervasives.Native.option Spec.Box.secretkey\nEffect: Prims.Tot\n\nContext:\nmodule Spec.Box\n\nopen FStar.Mul\nopen Lib.IntTypes\nopen Lib.Sequence\nopen Lib.ByteSequence\n\nopen Spec.SecretBox\n\n#set-options \"--z3rlimit 20 --max_fuel 0 --max_ifuel 0\"\n\n(* Constants *)\nlet size_publickey = 32 (* in bytes *)\nlet size_secretkey = 32 (* in bytes *)\n\ntype publickey = lbytes size_publickey\ntype secretkey = lbytes size_secretkey\n\nPartial implementation:\nlet ecdh (pk: publickey) (sk: secretkey) : option secretkey =\n \n"
},
{
"role": "assistant",
"content": "let shared = Spec.Curve25519.scalarmult sk pk in\nif not (lbytes_eq shared (create 32 (u8 0))) then Some shared else None"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: pk: Spec.Box.publickey -> sk: Spec.Box.secretkey\n -> FStar.Pervasives.Native.option Spec.SecretBox.key\nEffect: Prims.Tot\n\nContext:\nmodule Spec.Box\n\nopen FStar.Mul\nopen Lib.IntTypes\nopen Lib.Sequence\nopen Lib.ByteSequence\n\nopen Spec.SecretBox\n\n#set-options \"--z3rlimit 20 --max_fuel 0 --max_ifuel 0\"\n\n(* Constants *)\nlet size_publickey = 32 (* in bytes *)\nlet size_secretkey = 32 (* in bytes *)\n\ntype publickey = lbytes size_publickey\ntype secretkey = lbytes size_secretkey\n\n\nlet ecdh (pk:publickey) (sk:secretkey) : option secretkey =\n let shared = Spec.Curve25519.scalarmult sk pk in\n if not (lbytes_eq shared (create 32 (u8 0))) then\n Some shared\n else None\n\nPartial implementation:\nlet box_beforenm (pk: publickey) (sk: secretkey) : option key =\n \n"
},
{
"role": "assistant",
"content": "let shared = ecdh pk sk in\nmatch shared with\n| Some shared -> Some (Spec.Salsa20.hsalsa20 shared (create 16 (u8 0)))\n| None -> None"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n pk: Spec.Box.publickey ->\n sk: Spec.Box.secretkey ->\n n: Spec.SecretBox.nonce ->\n t: Spec.SecretBox.tag ->\n c:\n Lib.ByteSequence.bytes\n {Lib.Sequence.length c / Spec.SecretBox.size_block <= Lib.IntTypes.max_size_t}\n -> FStar.Pervasives.Native.option (m:\n Lib.ByteSequence.bytes{Lib.Sequence.length m = Lib.Sequence.length c})\nEffect: Prims.Tot\n\nContext:\nmodule Spec.Box\n\nopen FStar.Mul\nopen Lib.IntTypes\nopen Lib.Sequence\nopen Lib.ByteSequence\n\nopen Spec.SecretBox\n\n#set-options \"--z3rlimit 20 --max_fuel 0 --max_ifuel 0\"\n\n(* Constants *)\nlet size_publickey = 32 (* in bytes *)\nlet size_secretkey = 32 (* in bytes *)\n\ntype publickey = lbytes size_publickey\ntype secretkey = lbytes size_secretkey\n\n\nlet ecdh (pk:publickey) (sk:secretkey) : option secretkey =\n let shared = Spec.Curve25519.scalarmult sk pk in\n if not (lbytes_eq shared (create 32 (u8 0))) then\n Some shared\n else None\n\n\nlet box_beforenm (pk:publickey) (sk:secretkey) : option key =\n let shared = ecdh pk sk in\n match shared with\n | Some shared -> Some (Spec.Salsa20.hsalsa20 shared (create 16 (u8 0)))\n | None -> None\n\n\nlet box_detached_afternm (k:key) (n:nonce) (m:bytes{length m / size_block <= max_size_t}) : tag & c:bytes{length c = length m} =\n secretbox_detached k n m\n\n\nlet box_detached (sk:secretkey) (pk:publickey) (n:nonce) (m:bytes{length m / size_block <= max_size_t}) : option (tag & c:bytes{length c = length m}) =\n let k = box_beforenm pk sk in\n match k with\n | Some k -> Some (box_detached_afternm k n m)\n | None -> None\n\n\nlet box_open_detached_afternm (k:key) (n:nonce) (t:tag) (c:bytes{length c / size_block <= max_size_t}) : option (m:bytes{length m = length c}) =\n secretbox_open_detached k n t c\n\nPartial implementation:\nlet box_open_detached\n (pk: publickey)\n (sk: secretkey)\n (n: nonce)\n (t: tag)\n (c: bytes{length c / size_block <= max_size_t})\n : option (m: bytes{length m = length c}) =\n \n"
},
{
"role": "assistant",
"content": "let k = box_beforenm pk sk in\nmatch k with\n| Some k -> box_open_detached_afternm k n t c\n| None -> None"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n sk: Spec.Box.secretkey ->\n pk: Spec.Box.publickey ->\n n: Spec.SecretBox.nonce ->\n m:\n Lib.ByteSequence.bytes\n {Lib.Sequence.length m / Spec.SecretBox.size_block <= Lib.IntTypes.max_size_t}\n -> FStar.Pervasives.Native.option (Spec.SecretBox.tag *\n c:\n Lib.ByteSequence.bytes{Lib.Sequence.length c = Lib.Sequence.length m})\nEffect: Prims.Tot\n\nContext:\nmodule Spec.Box\n\nopen FStar.Mul\nopen Lib.IntTypes\nopen Lib.Sequence\nopen Lib.ByteSequence\n\nopen Spec.SecretBox\n\n#set-options \"--z3rlimit 20 --max_fuel 0 --max_ifuel 0\"\n\n(* Constants *)\nlet size_publickey = 32 (* in bytes *)\nlet size_secretkey = 32 (* in bytes *)\n\ntype publickey = lbytes size_publickey\ntype secretkey = lbytes size_secretkey\n\n\nlet ecdh (pk:publickey) (sk:secretkey) : option secretkey =\n let shared = Spec.Curve25519.scalarmult sk pk in\n if not (lbytes_eq shared (create 32 (u8 0))) then\n Some shared\n else None\n\n\nlet box_beforenm (pk:publickey) (sk:secretkey) : option key =\n let shared = ecdh pk sk in\n match shared with\n | Some shared -> Some (Spec.Salsa20.hsalsa20 shared (create 16 (u8 0)))\n | None -> None\n\n\nlet box_detached_afternm (k:key) (n:nonce) (m:bytes{length m / size_block <= max_size_t}) : tag & c:bytes{length c = length m} =\n secretbox_detached k n m\n\nPartial implementation:\nlet box_detached\n (sk: secretkey)\n (pk: publickey)\n (n: nonce)\n (m: bytes{length m / size_block <= max_size_t})\n : option (tag & c: bytes{length c = length m}) =\n \n"
},
{
"role": "assistant",
"content": "let k = box_beforenm pk sk in\nmatch k with\n| Some k -> Some (box_detached_afternm k n m)\n| None -> None"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: h: FStar.Monotonic.HyperStack.mem -> c: LowStar.Lib.LinkedList.t a -> l: Prims.list a\n -> Prims.Ghost (Prims.list (LowStar.Buffer.pointer (LowStar.Lib.LinkedList.cell a)))\nEffect: Prims.Ghost\n\nContext:\nmodule LowStar.Lib.LinkedList\n\n/// This module, oftentimes referred to as LL1 in other parts of the\n/// documentation, provides a low-level view over a linked list. Unless you\n/// intend on modifying the structure of the linked list (e.g. removing cells)\n/// or iterating over each cell, your are probably better off using LL2.\n///\n/// This module is intentionally not relying on a tight abstraction, to i) permit\n/// direct-style iteration by clients and ii) allow stating some SMTPat lemmas\n/// only on LL1 rather than replicating them in LL2 (going from LL2 to LL1 can\n/// happen automatically at the SMT level since LL2 does not have an fsti either).\n///\n/// Some open questions relative to the design of this module:\n/// - should ``v`` be defined directly in terms of ``cells`` rather than have a\n/// recursive definition along with an equivalence lemma?\n/// - should ``footprint`` be defined directly in terms of ``cells`` rather than\n/// have a recursive definition along with an equivalence lemma?\n/// - is it worth the trouble defining a footprint? my intuition is that it's\n/// fine and allows some clients to do very precise reasoning -- other modules\n/// (e.g. AssocList), when establishing abstraction boundaries, can simplify\n/// things and get rid of the precise footprint reasoning in favor of a static\n/// region-based footprint\n///\n/// Some remaining work items:\n/// - switch to C.Loops.while\n\nopen LowStar.BufferOps\n\nmodule B = LowStar.Buffer\nmodule HS = FStar.HyperStack\nmodule G = FStar.Ghost\nmodule L = FStar.List.Tot\nmodule U32 = FStar.UInt32\nmodule ST = FStar.HyperStack.ST\n\nopen FStar.HyperStack.ST\n\n/// Breaking from my own recommendation here... using fuel 1 / ifuel 1 since\n/// everything does recursive list reasoning.\n#set-options \"--__no_positivity --fuel 1 --ifuel 1\"\n\n/// Definition of the Low* type\n/// ---------------------------\nnoeq\ntype t (a: Type0) =\n B.pointer_or_null (cell a)\n\nand cell (a: Type0) = {\n next: t a;\n data: a;\n}\n\n/// Unlike the canonical style that associates a ``v`` function to any stateful\n/// representation, we use a relation here. Writing a function is just\n/// impossible, since we can't prove termination of ``v`` owing to the fact that\n/// there may be an infinite number of cells in the heap.\n///\n/// LL2 is the module that switches to the function style.\n///\n/// Note: no need to use erased here for ``l`` since the function is already in ``GTot``.\nlet rec well_formed #a (h: HS.mem) (c: t a) (l: list a): GTot Type0 (decreases l) =\n B.live h c /\\ (\n match l with\n | [] ->\n B.g_is_null c\n | a :: q ->\n B.length c == 1 /\\ (\n let { data=data; next=next } = B.get h c 0 in\n a == data /\\\n well_formed h next (G.hide q)\n ))\n\n/// Precise reasoning via the cells predicate\n/// -----------------------------------------\n///\n/// The modifies clauses are given either on the footprint of the list, or on a\n/// region that includes the footprint of the list (higher up the stack, e.g.\n/// LL2). This is good, but does not rule out \"stupid\" implementations that\n/// might re-allocate every single list cell when, say, doing a pop.\n///\n/// For advanced usages which take pointers directly to individual list cells,\n/// it's important to rule out these cases. We thus define the ``cells``\n/// predicate for that purpose.\n\nPartial implementation:\nlet rec cells #a (h: HS.mem) (c: t a) (l: list a)\n : Ghost (list (B.pointer (cell a)))\n (requires well_formed h c l)\n (ensures fun _ -> True)\n (decreases l) =\n \n"
},
{
"role": "assistant",
"content": "if B.g_is_null c then [] else c :: cells h (B.deref h c).next (List.Tot.tl l)"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: r: Steel.GhostPCMReference.ref a p -> v0: a -> v1: a\n -> Steel.Effect.Atomic.SteelGhostT (_: Prims.unit{FStar.PCM.composable p v0 v1})\nEffect: Steel.Effect.Atomic.SteelGhostT\n\nContext:\n(*\n Copyright 2020 Microsoft Research\n\n Licensed under the Apache License, Version 2.0 (the \"License\");\n you may not use this file except in compliance with the License.\n You may obtain a copy of the License at\n\n http://www.apache.org/licenses/LICENSE-2.0\n\n Unless required by applicable law or agreed to in writing, software\n distributed under the License is distributed on an \"AS IS\" BASIS,\n WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n See the License for the specific language governing permissions and\n limitations under the License.\n*)\nmodule Steel.GhostPCMReference\n(* A ghost variant of Steel.PCMReference *)\nopen FStar.PCM\nopen FStar.Ghost\n\nopen Steel.Memory\nopen Steel.Effect.Atomic\nopen Steel.Effect\nmodule Mem = Steel.Memory\nmodule P = Steel.PCMReference\n\nlet ref (a:Type) (p:pcm a) = erased (Steel.Memory.ref a p)\n\n/// Its selector is non-informative (it is unit)\n[@@__reduce__]\nlet pts_to (#a:Type u#1) (#pcm:pcm a) (r:ref a pcm) ([@@@smt_fallback]v:a)\n = to_vprop (Steel.Memory.pts_to r v)\n\nlet alloc (#o:inames)\n (#a:Type)\n (#pcm:pcm a)\n (x:a)\n : SteelGhost\n (ref a pcm) o\n (emp)\n (fun r -> pts_to r x)\n (requires fun _ -> pcm.refine x)\n (ensures fun _ _ _ -> True)\n = rewrite_slprop emp (to_vprop Mem.emp) (fun _ -> reveal_emp ());\n FStar.PCM.compatible_refl pcm x;\n let r = as_atomic_action_ghost (alloc_action o x) in\n r\n\nlet read (#o:inames)\n (#a:Type)\n (#pcm:pcm a)\n (#v0:a)\n (r:ref a pcm)\n : SteelGhost a o\n (pts_to r v0)\n (fun _ -> pts_to r v0)\n (requires fun _ -> True)\n (ensures fun _ v _ -> compatible pcm v0 v)\n = let v = as_atomic_action_ghost (sel_action o r v0) in\n v\n\nlet write (#o:inames)\n (#a:Type)\n (#pcm:pcm a)\n (r:ref a pcm)\n (v0:a)\n (v1:a)\n : SteelGhost unit o\n (pts_to r v0)\n (fun _ -> pts_to r v1)\n (requires fun _ -> frame_preserving pcm v0 v1 /\\ pcm.refine v1)\n (ensures fun _ _ _ -> True)\n = as_atomic_action_ghost (upd_action o r v0 v1)\n\nlet upd_gen (#o:inames)\n (#a:Type)\n (#p:pcm a)\n (r:ref a p)\n (x y:a)\n (f:frame_preserving_upd p x y)\n : SteelGhostT unit o\n (pts_to r x)\n (fun _ -> pts_to r y)\n = as_atomic_action_ghost (Steel.Memory.upd_gen o r x y f)\n\nlet share (#o:inames)\n (#a:Type)\n (#p:pcm a)\n (r:ref a p)\n (v:a)\n (v0:a)\n (v1:a)\n : SteelGhost unit o\n (pts_to r v)\n (fun _ -> pts_to r v0 `star` pts_to r v1)\n (requires fun _ ->\n composable p v0 v1 /\\\n v == op p v0 v1)\n (ensures fun _ _ _ -> True)\n = P.split r v v0 v1\n\nPartial implementation:\nlet gather (#o: inames) (#a: Type) (#p: FStar.PCM.pcm a) (r: ref a p) (v0 v1: a)\n : SteelGhostT (_: unit{composable p v0 v1})\n o\n ((pts_to r v0) `star` (pts_to r v1))\n (fun _ -> pts_to r (op p v0 v1)) =\n \n"
},
{
"role": "assistant",
"content": "P.gather r v0 v1"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: h: FStar.Monotonic.HyperStack.mem -> l: LowStar.Lib.LinkedList.t a -> n: Prims.list a\n -> Prims.Ghost LowStar.Monotonic.Buffer.loc\nEffect: Prims.Ghost\n\nContext:\nmodule LowStar.Lib.LinkedList\n\n/// This module, oftentimes referred to as LL1 in other parts of the\n/// documentation, provides a low-level view over a linked list. Unless you\n/// intend on modifying the structure of the linked list (e.g. removing cells)\n/// or iterating over each cell, your are probably better off using LL2.\n///\n/// This module is intentionally not relying on a tight abstraction, to i) permit\n/// direct-style iteration by clients and ii) allow stating some SMTPat lemmas\n/// only on LL1 rather than replicating them in LL2 (going from LL2 to LL1 can\n/// happen automatically at the SMT level since LL2 does not have an fsti either).\n///\n/// Some open questions relative to the design of this module:\n/// - should ``v`` be defined directly in terms of ``cells`` rather than have a\n/// recursive definition along with an equivalence lemma?\n/// - should ``footprint`` be defined directly in terms of ``cells`` rather than\n/// have a recursive definition along with an equivalence lemma?\n/// - is it worth the trouble defining a footprint? my intuition is that it's\n/// fine and allows some clients to do very precise reasoning -- other modules\n/// (e.g. AssocList), when establishing abstraction boundaries, can simplify\n/// things and get rid of the precise footprint reasoning in favor of a static\n/// region-based footprint\n///\n/// Some remaining work items:\n/// - switch to C.Loops.while\n\nopen LowStar.BufferOps\n\nmodule B = LowStar.Buffer\nmodule HS = FStar.HyperStack\nmodule G = FStar.Ghost\nmodule L = FStar.List.Tot\nmodule U32 = FStar.UInt32\nmodule ST = FStar.HyperStack.ST\n\nopen FStar.HyperStack.ST\n\n/// Breaking from my own recommendation here... using fuel 1 / ifuel 1 since\n/// everything does recursive list reasoning.\n#set-options \"--__no_positivity --fuel 1 --ifuel 1\"\n\n/// Definition of the Low* type\n/// ---------------------------\nnoeq\ntype t (a: Type0) =\n B.pointer_or_null (cell a)\n\nand cell (a: Type0) = {\n next: t a;\n data: a;\n}\n\n/// Unlike the canonical style that associates a ``v`` function to any stateful\n/// representation, we use a relation here. Writing a function is just\n/// impossible, since we can't prove termination of ``v`` owing to the fact that\n/// there may be an infinite number of cells in the heap.\n///\n/// LL2 is the module that switches to the function style.\n///\n/// Note: no need to use erased here for ``l`` since the function is already in ``GTot``.\nlet rec well_formed #a (h: HS.mem) (c: t a) (l: list a): GTot Type0 (decreases l) =\n B.live h c /\\ (\n match l with\n | [] ->\n B.g_is_null c\n | a :: q ->\n B.length c == 1 /\\ (\n let { data=data; next=next } = B.get h c 0 in\n a == data /\\\n well_formed h next (G.hide q)\n ))\n\n/// Precise reasoning via the cells predicate\n/// -----------------------------------------\n///\n/// The modifies clauses are given either on the footprint of the list, or on a\n/// region that includes the footprint of the list (higher up the stack, e.g.\n/// LL2). This is good, but does not rule out \"stupid\" implementations that\n/// might re-allocate every single list cell when, say, doing a pop.\n///\n/// For advanced usages which take pointers directly to individual list cells,\n/// it's important to rule out these cases. We thus define the ``cells``\n/// predicate for that purpose.\n\nlet rec cells #a (h: HS.mem) (c: t a) (l: list a): Ghost (list (B.pointer (cell a)))\n (requires well_formed h c l)\n (ensures fun _ -> True)\n (decreases l)\n=\n if B.g_is_null c then\n []\n else\n c :: cells h (B.deref h c).next (List.Tot.tl l)\n\n/// A missing bit of automated reasoning\n/// ------------------------------------\n\n/// This should be absolutely trivial, yet, because there's no pattern on\n/// null_unique, we have to do the case analysis and the decomposition\n/// ourselves. This is quite tedious, as it usually takes a while (at least it\n/// did for me!) to realize that this property does not hold automatically.\n///\n/// I'm setting an SMTPat on this lemma because it's absolutely impossible to\n/// get any work done without it.\nlet same_cells_same_pointer #a (h0 h1: HS.mem) (ll0 ll1: t a) (l0 l1: list a): Lemma\n (requires (\n well_formed h0 ll0 l0 /\\\n well_formed h1 ll1 l1 /\\\n cells h0 ll0 l0 == cells h1 ll1 l1))\n (ensures (\n ll0 == ll1))\n [ SMTPat (cells h0 ll0 l0); SMTPat (cells h1 ll1 l1) ]\n=\n if B.g_is_null ll0 && B.g_is_null ll1 then begin\n B.null_unique ll0;\n B.null_unique ll1\n end else if not (B.g_is_null ll0) && not (B.g_is_null ll1) then begin\n ()\n end else\n false_elim ()\n\n/// Classic stateful reasoning lemmas\n/// ---------------------------------\n///\n/// This essentially follows ``stateful``, the type class of a mutable piece of\n/// state defined using the modifies-theory of Low*v2 and equipped with the\n/// right lemmas to enable usage by clients. See\n/// hacl-star/code/streaming/Hacl.Streaming.Interface.fst for more context.\n\nval footprint: (#a: Type) -> (h: HS.mem) -> (l: t a) -> (n: list a) -> Ghost B.loc\n (requires (well_formed h l n))\n (ensures (fun refs -> True))\n (decreases n)\n\nPartial implementation:\nlet rec footprint #a h l n =\n \n"
},
{
"role": "assistant",
"content": "if B.g_is_null l\nthen B.loc_none\nelse\n let { next = next } = B.get h l 0 in\n let refs = footprint h next (G.hide (L.tl n)) in\n B.loc_union (B.loc_addr_of_buffer l) refs"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n k: Spec.SecretBox.key ->\n n: Spec.SecretBox.nonce ->\n m:\n Lib.ByteSequence.bytes\n {Lib.Sequence.length m / Spec.SecretBox.size_block <= Lib.IntTypes.max_size_t}\n -> Spec.SecretBox.tag * c: Lib.ByteSequence.bytes{Lib.Sequence.length c = Lib.Sequence.length m}\nEffect: Prims.Tot\n\nContext:\nmodule Spec.Box\n\nopen FStar.Mul\nopen Lib.IntTypes\nopen Lib.Sequence\nopen Lib.ByteSequence\n\nopen Spec.SecretBox\n\n#set-options \"--z3rlimit 20 --max_fuel 0 --max_ifuel 0\"\n\n(* Constants *)\nlet size_publickey = 32 (* in bytes *)\nlet size_secretkey = 32 (* in bytes *)\n\ntype publickey = lbytes size_publickey\ntype secretkey = lbytes size_secretkey\n\n\nlet ecdh (pk:publickey) (sk:secretkey) : option secretkey =\n let shared = Spec.Curve25519.scalarmult sk pk in\n if not (lbytes_eq shared (create 32 (u8 0))) then\n Some shared\n else None\n\n\nlet box_beforenm (pk:publickey) (sk:secretkey) : option key =\n let shared = ecdh pk sk in\n match shared with\n | Some shared -> Some (Spec.Salsa20.hsalsa20 shared (create 16 (u8 0)))\n | None -> None\n\nPartial implementation:\nlet box_detached_afternm (k: key) (n: nonce) (m: bytes{length m / size_block <= max_size_t})\n : tag & c: bytes{length c = length m} =\n \n"
},
{
"role": "assistant",
"content": "secretbox_detached k n m"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: h: FStar.Monotonic.HyperStack.mem -> c: LowStar.Lib.LinkedList.t a -> l: Prims.list a\n -> Prims.GTot Type0\nEffect: Prims.GTot\n\nContext:\nmodule LowStar.Lib.LinkedList\n\n/// This module, oftentimes referred to as LL1 in other parts of the\n/// documentation, provides a low-level view over a linked list. Unless you\n/// intend on modifying the structure of the linked list (e.g. removing cells)\n/// or iterating over each cell, your are probably better off using LL2.\n///\n/// This module is intentionally not relying on a tight abstraction, to i) permit\n/// direct-style iteration by clients and ii) allow stating some SMTPat lemmas\n/// only on LL1 rather than replicating them in LL2 (going from LL2 to LL1 can\n/// happen automatically at the SMT level since LL2 does not have an fsti either).\n///\n/// Some open questions relative to the design of this module:\n/// - should ``v`` be defined directly in terms of ``cells`` rather than have a\n/// recursive definition along with an equivalence lemma?\n/// - should ``footprint`` be defined directly in terms of ``cells`` rather than\n/// have a recursive definition along with an equivalence lemma?\n/// - is it worth the trouble defining a footprint? my intuition is that it's\n/// fine and allows some clients to do very precise reasoning -- other modules\n/// (e.g. AssocList), when establishing abstraction boundaries, can simplify\n/// things and get rid of the precise footprint reasoning in favor of a static\n/// region-based footprint\n///\n/// Some remaining work items:\n/// - switch to C.Loops.while\n\nopen LowStar.BufferOps\n\nmodule B = LowStar.Buffer\nmodule HS = FStar.HyperStack\nmodule G = FStar.Ghost\nmodule L = FStar.List.Tot\nmodule U32 = FStar.UInt32\nmodule ST = FStar.HyperStack.ST\n\nopen FStar.HyperStack.ST\n\n/// Breaking from my own recommendation here... using fuel 1 / ifuel 1 since\n/// everything does recursive list reasoning.\n#set-options \"--__no_positivity --fuel 1 --ifuel 1\"\n\n/// Definition of the Low* type\n/// ---------------------------\nnoeq\ntype t (a: Type0) =\n B.pointer_or_null (cell a)\n\nand cell (a: Type0) = {\n next: t a;\n data: a;\n}\n\n/// Unlike the canonical style that associates a ``v`` function to any stateful\n/// representation, we use a relation here. Writing a function is just\n/// impossible, since we can't prove termination of ``v`` owing to the fact that\n/// there may be an infinite number of cells in the heap.\n///\n/// LL2 is the module that switches to the function style.\n///\n\nPartial implementation:\nlet rec well_formed #a (h: HS.mem) (c: t a) (l: list a) : GTot Type0 (decreases l) =\n \n"
},
{
"role": "assistant",
"content": "B.live h c /\\\n(match l with\n | [] -> B.g_is_null c\n | a :: q ->\n B.length c == 1 /\\\n (let { data = data ; next = next } = B.get h c 0 in\n a == data /\\ well_formed h next (G.hide q)))"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: h: FStar.Monotonic.HyperStack.mem -> l: LowStar.Lib.LinkedList.t a -> n: Prims.list a\n -> Prims.Ghost Type0\nEffect: Prims.Ghost\n\nContext:\nmodule LowStar.Lib.LinkedList\n\n/// This module, oftentimes referred to as LL1 in other parts of the\n/// documentation, provides a low-level view over a linked list. Unless you\n/// intend on modifying the structure of the linked list (e.g. removing cells)\n/// or iterating over each cell, your are probably better off using LL2.\n///\n/// This module is intentionally not relying on a tight abstraction, to i) permit\n/// direct-style iteration by clients and ii) allow stating some SMTPat lemmas\n/// only on LL1 rather than replicating them in LL2 (going from LL2 to LL1 can\n/// happen automatically at the SMT level since LL2 does not have an fsti either).\n///\n/// Some open questions relative to the design of this module:\n/// - should ``v`` be defined directly in terms of ``cells`` rather than have a\n/// recursive definition along with an equivalence lemma?\n/// - should ``footprint`` be defined directly in terms of ``cells`` rather than\n/// have a recursive definition along with an equivalence lemma?\n/// - is it worth the trouble defining a footprint? my intuition is that it's\n/// fine and allows some clients to do very precise reasoning -- other modules\n/// (e.g. AssocList), when establishing abstraction boundaries, can simplify\n/// things and get rid of the precise footprint reasoning in favor of a static\n/// region-based footprint\n///\n/// Some remaining work items:\n/// - switch to C.Loops.while\n\nopen LowStar.BufferOps\n\nmodule B = LowStar.Buffer\nmodule HS = FStar.HyperStack\nmodule G = FStar.Ghost\nmodule L = FStar.List.Tot\nmodule U32 = FStar.UInt32\nmodule ST = FStar.HyperStack.ST\n\nopen FStar.HyperStack.ST\n\n/// Breaking from my own recommendation here... using fuel 1 / ifuel 1 since\n/// everything does recursive list reasoning.\n#set-options \"--__no_positivity --fuel 1 --ifuel 1\"\n\n/// Definition of the Low* type\n/// ---------------------------\nnoeq\ntype t (a: Type0) =\n B.pointer_or_null (cell a)\n\nand cell (a: Type0) = {\n next: t a;\n data: a;\n}\n\n/// Unlike the canonical style that associates a ``v`` function to any stateful\n/// representation, we use a relation here. Writing a function is just\n/// impossible, since we can't prove termination of ``v`` owing to the fact that\n/// there may be an infinite number of cells in the heap.\n///\n/// LL2 is the module that switches to the function style.\n///\n/// Note: no need to use erased here for ``l`` since the function is already in ``GTot``.\nlet rec well_formed #a (h: HS.mem) (c: t a) (l: list a): GTot Type0 (decreases l) =\n B.live h c /\\ (\n match l with\n | [] ->\n B.g_is_null c\n | a :: q ->\n B.length c == 1 /\\ (\n let { data=data; next=next } = B.get h c 0 in\n a == data /\\\n well_formed h next (G.hide q)\n ))\n\n/// Precise reasoning via the cells predicate\n/// -----------------------------------------\n///\n/// The modifies clauses are given either on the footprint of the list, or on a\n/// region that includes the footprint of the list (higher up the stack, e.g.\n/// LL2). This is good, but does not rule out \"stupid\" implementations that\n/// might re-allocate every single list cell when, say, doing a pop.\n///\n/// For advanced usages which take pointers directly to individual list cells,\n/// it's important to rule out these cases. We thus define the ``cells``\n/// predicate for that purpose.\n\nlet rec cells #a (h: HS.mem) (c: t a) (l: list a): Ghost (list (B.pointer (cell a)))\n (requires well_formed h c l)\n (ensures fun _ -> True)\n (decreases l)\n=\n if B.g_is_null c then\n []\n else\n c :: cells h (B.deref h c).next (List.Tot.tl l)\n\n/// A missing bit of automated reasoning\n/// ------------------------------------\n\n/// This should be absolutely trivial, yet, because there's no pattern on\n/// null_unique, we have to do the case analysis and the decomposition\n/// ourselves. This is quite tedious, as it usually takes a while (at least it\n/// did for me!) to realize that this property does not hold automatically.\n///\n/// I'm setting an SMTPat on this lemma because it's absolutely impossible to\n/// get any work done without it.\nlet same_cells_same_pointer #a (h0 h1: HS.mem) (ll0 ll1: t a) (l0 l1: list a): Lemma\n (requires (\n well_formed h0 ll0 l0 /\\\n well_formed h1 ll1 l1 /\\\n cells h0 ll0 l0 == cells h1 ll1 l1))\n (ensures (\n ll0 == ll1))\n [ SMTPat (cells h0 ll0 l0); SMTPat (cells h1 ll1 l1) ]\n=\n if B.g_is_null ll0 && B.g_is_null ll1 then begin\n B.null_unique ll0;\n B.null_unique ll1\n end else if not (B.g_is_null ll0) && not (B.g_is_null ll1) then begin\n ()\n end else\n false_elim ()\n\n/// Classic stateful reasoning lemmas\n/// ---------------------------------\n///\n/// This essentially follows ``stateful``, the type class of a mutable piece of\n/// state defined using the modifies-theory of Low*v2 and equipped with the\n/// right lemmas to enable usage by clients. See\n/// hacl-star/code/streaming/Hacl.Streaming.Interface.fst for more context.\n\nval footprint: (#a: Type) -> (h: HS.mem) -> (l: t a) -> (n: list a) -> Ghost B.loc\n (requires (well_formed h l n))\n (ensures (fun refs -> True))\n (decreases n)\n\nlet rec footprint #a h l n =\n if B.g_is_null l\n then B.loc_none\n else\n let {next = next} = B.get h l 0 in\n let refs = footprint h next (G.hide (L.tl n)) in\n B.loc_union (B.loc_addr_of_buffer l) refs\n\n/// Departing from LinkedList4 here. I prefer to bolt these into the invariant\n/// rather than requiring on some wizard lemmas from LowStar.Monotonic.Buffer to\n/// deduce these.\nval cells_pairwise_disjoint: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nPartial implementation:\nlet rec cells_pairwise_disjoint #_ h l n =\n \n"
},
{
"role": "assistant",
"content": "if B.g_is_null l\nthen True\nelse\n let next = (B.deref h l).next in\n B.loc_disjoint (B.loc_addr_of_buffer l) (footprint h next (L.tl n)) /\\\n cells_pairwise_disjoint h next (L.tl n)"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: h: FStar.Monotonic.HyperStack.mem -> l: LowStar.Lib.LinkedList.t a -> n: Prims.list a\n -> Prims.Ghost Type0\nEffect: Prims.Ghost\n\nContext:\nmodule LowStar.Lib.LinkedList\n\n/// This module, oftentimes referred to as LL1 in other parts of the\n/// documentation, provides a low-level view over a linked list. Unless you\n/// intend on modifying the structure of the linked list (e.g. removing cells)\n/// or iterating over each cell, your are probably better off using LL2.\n///\n/// This module is intentionally not relying on a tight abstraction, to i) permit\n/// direct-style iteration by clients and ii) allow stating some SMTPat lemmas\n/// only on LL1 rather than replicating them in LL2 (going from LL2 to LL1 can\n/// happen automatically at the SMT level since LL2 does not have an fsti either).\n///\n/// Some open questions relative to the design of this module:\n/// - should ``v`` be defined directly in terms of ``cells`` rather than have a\n/// recursive definition along with an equivalence lemma?\n/// - should ``footprint`` be defined directly in terms of ``cells`` rather than\n/// have a recursive definition along with an equivalence lemma?\n/// - is it worth the trouble defining a footprint? my intuition is that it's\n/// fine and allows some clients to do very precise reasoning -- other modules\n/// (e.g. AssocList), when establishing abstraction boundaries, can simplify\n/// things and get rid of the precise footprint reasoning in favor of a static\n/// region-based footprint\n///\n/// Some remaining work items:\n/// - switch to C.Loops.while\n\nopen LowStar.BufferOps\n\nmodule B = LowStar.Buffer\nmodule HS = FStar.HyperStack\nmodule G = FStar.Ghost\nmodule L = FStar.List.Tot\nmodule U32 = FStar.UInt32\nmodule ST = FStar.HyperStack.ST\n\nopen FStar.HyperStack.ST\n\n/// Breaking from my own recommendation here... using fuel 1 / ifuel 1 since\n/// everything does recursive list reasoning.\n#set-options \"--__no_positivity --fuel 1 --ifuel 1\"\n\n/// Definition of the Low* type\n/// ---------------------------\nnoeq\ntype t (a: Type0) =\n B.pointer_or_null (cell a)\n\nand cell (a: Type0) = {\n next: t a;\n data: a;\n}\n\n/// Unlike the canonical style that associates a ``v`` function to any stateful\n/// representation, we use a relation here. Writing a function is just\n/// impossible, since we can't prove termination of ``v`` owing to the fact that\n/// there may be an infinite number of cells in the heap.\n///\n/// LL2 is the module that switches to the function style.\n///\n/// Note: no need to use erased here for ``l`` since the function is already in ``GTot``.\nlet rec well_formed #a (h: HS.mem) (c: t a) (l: list a): GTot Type0 (decreases l) =\n B.live h c /\\ (\n match l with\n | [] ->\n B.g_is_null c\n | a :: q ->\n B.length c == 1 /\\ (\n let { data=data; next=next } = B.get h c 0 in\n a == data /\\\n well_formed h next (G.hide q)\n ))\n\n/// Precise reasoning via the cells predicate\n/// -----------------------------------------\n///\n/// The modifies clauses are given either on the footprint of the list, or on a\n/// region that includes the footprint of the list (higher up the stack, e.g.\n/// LL2). This is good, but does not rule out \"stupid\" implementations that\n/// might re-allocate every single list cell when, say, doing a pop.\n///\n/// For advanced usages which take pointers directly to individual list cells,\n/// it's important to rule out these cases. We thus define the ``cells``\n/// predicate for that purpose.\n\nlet rec cells #a (h: HS.mem) (c: t a) (l: list a): Ghost (list (B.pointer (cell a)))\n (requires well_formed h c l)\n (ensures fun _ -> True)\n (decreases l)\n=\n if B.g_is_null c then\n []\n else\n c :: cells h (B.deref h c).next (List.Tot.tl l)\n\n/// A missing bit of automated reasoning\n/// ------------------------------------\n\n/// This should be absolutely trivial, yet, because there's no pattern on\n/// null_unique, we have to do the case analysis and the decomposition\n/// ourselves. This is quite tedious, as it usually takes a while (at least it\n/// did for me!) to realize that this property does not hold automatically.\n///\n/// I'm setting an SMTPat on this lemma because it's absolutely impossible to\n/// get any work done without it.\nlet same_cells_same_pointer #a (h0 h1: HS.mem) (ll0 ll1: t a) (l0 l1: list a): Lemma\n (requires (\n well_formed h0 ll0 l0 /\\\n well_formed h1 ll1 l1 /\\\n cells h0 ll0 l0 == cells h1 ll1 l1))\n (ensures (\n ll0 == ll1))\n [ SMTPat (cells h0 ll0 l0); SMTPat (cells h1 ll1 l1) ]\n=\n if B.g_is_null ll0 && B.g_is_null ll1 then begin\n B.null_unique ll0;\n B.null_unique ll1\n end else if not (B.g_is_null ll0) && not (B.g_is_null ll1) then begin\n ()\n end else\n false_elim ()\n\n/// Classic stateful reasoning lemmas\n/// ---------------------------------\n///\n/// This essentially follows ``stateful``, the type class of a mutable piece of\n/// state defined using the modifies-theory of Low*v2 and equipped with the\n/// right lemmas to enable usage by clients. See\n/// hacl-star/code/streaming/Hacl.Streaming.Interface.fst for more context.\n\nval footprint: (#a: Type) -> (h: HS.mem) -> (l: t a) -> (n: list a) -> Ghost B.loc\n (requires (well_formed h l n))\n (ensures (fun refs -> True))\n (decreases n)\n\nlet rec footprint #a h l n =\n if B.g_is_null l\n then B.loc_none\n else\n let {next = next} = B.get h l 0 in\n let refs = footprint h next (G.hide (L.tl n)) in\n B.loc_union (B.loc_addr_of_buffer l) refs\n\n/// Departing from LinkedList4 here. I prefer to bolt these into the invariant\n/// rather than requiring on some wizard lemmas from LowStar.Monotonic.Buffer to\n/// deduce these.\nval cells_pairwise_disjoint: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_pairwise_disjoint #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.loc_disjoint (B.loc_addr_of_buffer l) (footprint h next (L.tl n)) /\\\n cells_pairwise_disjoint h next (L.tl n)\n\nval cells_live_freeable: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_live_freeable #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.live h l /\\ B.freeable l /\\\n cells_live_freeable h next (L.tl n)\n\nval invariant: #a:Type -> h:HS.mem -> l: t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n\nPartial implementation:\nlet invariant #a h l n =\n \n"
},
{
"role": "assistant",
"content": "cells_live_freeable h l n /\\ cells_pairwise_disjoint h l n"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n k: Spec.SecretBox.key ->\n n: Spec.SecretBox.nonce ->\n t: Spec.SecretBox.tag ->\n c:\n Lib.ByteSequence.bytes\n {Lib.Sequence.length c / Spec.SecretBox.size_block <= Lib.IntTypes.max_size_t}\n -> FStar.Pervasives.Native.option (m:\n Lib.ByteSequence.bytes{Lib.Sequence.length m = Lib.Sequence.length c})\nEffect: Prims.Tot\n\nContext:\nmodule Spec.Box\n\nopen FStar.Mul\nopen Lib.IntTypes\nopen Lib.Sequence\nopen Lib.ByteSequence\n\nopen Spec.SecretBox\n\n#set-options \"--z3rlimit 20 --max_fuel 0 --max_ifuel 0\"\n\n(* Constants *)\nlet size_publickey = 32 (* in bytes *)\nlet size_secretkey = 32 (* in bytes *)\n\ntype publickey = lbytes size_publickey\ntype secretkey = lbytes size_secretkey\n\n\nlet ecdh (pk:publickey) (sk:secretkey) : option secretkey =\n let shared = Spec.Curve25519.scalarmult sk pk in\n if not (lbytes_eq shared (create 32 (u8 0))) then\n Some shared\n else None\n\n\nlet box_beforenm (pk:publickey) (sk:secretkey) : option key =\n let shared = ecdh pk sk in\n match shared with\n | Some shared -> Some (Spec.Salsa20.hsalsa20 shared (create 16 (u8 0)))\n | None -> None\n\n\nlet box_detached_afternm (k:key) (n:nonce) (m:bytes{length m / size_block <= max_size_t}) : tag & c:bytes{length c = length m} =\n secretbox_detached k n m\n\n\nlet box_detached (sk:secretkey) (pk:publickey) (n:nonce) (m:bytes{length m / size_block <= max_size_t}) : option (tag & c:bytes{length c = length m}) =\n let k = box_beforenm pk sk in\n match k with\n | Some k -> Some (box_detached_afternm k n m)\n | None -> None\n\nPartial implementation:\nlet box_open_detached_afternm\n (k: key)\n (n: nonce)\n (t: tag)\n (c: bytes{length c / size_block <= max_size_t})\n : option (m: bytes{length m = length c}) =\n \n"
},
{
"role": "assistant",
"content": "secretbox_open_detached k n t c"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n k: Spec.SecretBox.key ->\n n: Spec.SecretBox.nonce ->\n m:\n Lib.ByteSequence.bytes\n {Lib.Sequence.length m / Spec.SecretBox.size_block <= Lib.IntTypes.max_size_t}\n -> c:\n Lib.ByteSequence.bytes{Lib.Sequence.length c = Spec.SecretBox.size_tag + Lib.Sequence.length m}\nEffect: Prims.Tot\n\nContext:\nmodule Spec.Box\n\nopen FStar.Mul\nopen Lib.IntTypes\nopen Lib.Sequence\nopen Lib.ByteSequence\n\nopen Spec.SecretBox\n\n#set-options \"--z3rlimit 20 --max_fuel 0 --max_ifuel 0\"\n\n(* Constants *)\nlet size_publickey = 32 (* in bytes *)\nlet size_secretkey = 32 (* in bytes *)\n\ntype publickey = lbytes size_publickey\ntype secretkey = lbytes size_secretkey\n\n\nlet ecdh (pk:publickey) (sk:secretkey) : option secretkey =\n let shared = Spec.Curve25519.scalarmult sk pk in\n if not (lbytes_eq shared (create 32 (u8 0))) then\n Some shared\n else None\n\n\nlet box_beforenm (pk:publickey) (sk:secretkey) : option key =\n let shared = ecdh pk sk in\n match shared with\n | Some shared -> Some (Spec.Salsa20.hsalsa20 shared (create 16 (u8 0)))\n | None -> None\n\n\nlet box_detached_afternm (k:key) (n:nonce) (m:bytes{length m / size_block <= max_size_t}) : tag & c:bytes{length c = length m} =\n secretbox_detached k n m\n\n\nlet box_detached (sk:secretkey) (pk:publickey) (n:nonce) (m:bytes{length m / size_block <= max_size_t}) : option (tag & c:bytes{length c = length m}) =\n let k = box_beforenm pk sk in\n match k with\n | Some k -> Some (box_detached_afternm k n m)\n | None -> None\n\n\nlet box_open_detached_afternm (k:key) (n:nonce) (t:tag) (c:bytes{length c / size_block <= max_size_t}) : option (m:bytes{length m = length c}) =\n secretbox_open_detached k n t c\n\n\nlet box_open_detached (pk:publickey) (sk:secretkey) (n:nonce) (t:tag) (c:bytes{length c / size_block <= max_size_t}) : option (m:bytes{length m = length c}) =\n let k = box_beforenm pk sk in\n match k with\n | Some k -> box_open_detached_afternm k n t c\n | None -> None\n\nPartial implementation:\nlet box_easy_afternm (k: key) (n: nonce) (m: bytes{length m / size_block <= max_size_t})\n : c: bytes{length c = size_tag + length m} =\n \n"
},
{
"role": "assistant",
"content": "let tg, c = box_detached_afternm k n m in\nSeq.append tg c"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n k: Spec.SecretBox.key ->\n n: Spec.SecretBox.nonce ->\n c:\n Lib.ByteSequence.bytes\n { Lib.Sequence.length c >= Spec.SecretBox.size_tag /\\\n (Lib.Sequence.length c - Spec.SecretBox.size_tag) / Spec.SecretBox.size_block <=\n Lib.IntTypes.max_size_t }\n -> FStar.Pervasives.Native.option (m:\n Lib.ByteSequence.bytes\n {Lib.Sequence.length m = Lib.Sequence.length c - Spec.SecretBox.size_tag})\nEffect: Prims.Tot\n\nContext:\nmodule Spec.Box\n\nopen FStar.Mul\nopen Lib.IntTypes\nopen Lib.Sequence\nopen Lib.ByteSequence\n\nopen Spec.SecretBox\n\n#set-options \"--z3rlimit 20 --max_fuel 0 --max_ifuel 0\"\n\n(* Constants *)\nlet size_publickey = 32 (* in bytes *)\nlet size_secretkey = 32 (* in bytes *)\n\ntype publickey = lbytes size_publickey\ntype secretkey = lbytes size_secretkey\n\n\nlet ecdh (pk:publickey) (sk:secretkey) : option secretkey =\n let shared = Spec.Curve25519.scalarmult sk pk in\n if not (lbytes_eq shared (create 32 (u8 0))) then\n Some shared\n else None\n\n\nlet box_beforenm (pk:publickey) (sk:secretkey) : option key =\n let shared = ecdh pk sk in\n match shared with\n | Some shared -> Some (Spec.Salsa20.hsalsa20 shared (create 16 (u8 0)))\n | None -> None\n\n\nlet box_detached_afternm (k:key) (n:nonce) (m:bytes{length m / size_block <= max_size_t}) : tag & c:bytes{length c = length m} =\n secretbox_detached k n m\n\n\nlet box_detached (sk:secretkey) (pk:publickey) (n:nonce) (m:bytes{length m / size_block <= max_size_t}) : option (tag & c:bytes{length c = length m}) =\n let k = box_beforenm pk sk in\n match k with\n | Some k -> Some (box_detached_afternm k n m)\n | None -> None\n\n\nlet box_open_detached_afternm (k:key) (n:nonce) (t:tag) (c:bytes{length c / size_block <= max_size_t}) : option (m:bytes{length m = length c}) =\n secretbox_open_detached k n t c\n\n\nlet box_open_detached (pk:publickey) (sk:secretkey) (n:nonce) (t:tag) (c:bytes{length c / size_block <= max_size_t}) : option (m:bytes{length m = length c}) =\n let k = box_beforenm pk sk in\n match k with\n | Some k -> box_open_detached_afternm k n t c\n | None -> None\n\n\nlet box_easy_afternm (k:key) (n:nonce) (m:bytes{length m / size_block <= max_size_t}) : c:bytes{length c = size_tag + length m} =\n let (tg, c) = box_detached_afternm k n m in\n Seq.append tg c\n\n\nlet box_easy (sk:secretkey) (pk:publickey) (n:nonce) (m:bytes{length m / size_block <= max_size_t}) : option (c:bytes{length c = size_tag + length m}) =\n let r = box_detached sk pk n m in\n match r with\n | Some (tg, c) -> Some (Seq.append tg c)\n | None -> None\n\nPartial implementation:\nlet box_open_easy_afternm\n (k: key)\n (n: nonce)\n (c: bytes{length c >= size_tag /\\ (length c - size_tag) / size_block <= max_size_t})\n : option (m: bytes{length m = length c - size_tag}) =\n \n"
},
{
"role": "assistant",
"content": "let tg = Seq.slice c 0 size_tag in\nlet e = Seq.slice c size_tag (length c) in\nbox_open_detached_afternm k n tg e"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n sk: Spec.Box.secretkey ->\n pk: Spec.Box.publickey ->\n n: Spec.SecretBox.nonce ->\n m:\n Lib.ByteSequence.bytes\n {Lib.Sequence.length m / Spec.SecretBox.size_block <= Lib.IntTypes.max_size_t}\n -> FStar.Pervasives.Native.option (c:\n Lib.ByteSequence.bytes\n {Lib.Sequence.length c = Spec.SecretBox.size_tag + Lib.Sequence.length m})\nEffect: Prims.Tot\n\nContext:\nmodule Spec.Box\n\nopen FStar.Mul\nopen Lib.IntTypes\nopen Lib.Sequence\nopen Lib.ByteSequence\n\nopen Spec.SecretBox\n\n#set-options \"--z3rlimit 20 --max_fuel 0 --max_ifuel 0\"\n\n(* Constants *)\nlet size_publickey = 32 (* in bytes *)\nlet size_secretkey = 32 (* in bytes *)\n\ntype publickey = lbytes size_publickey\ntype secretkey = lbytes size_secretkey\n\n\nlet ecdh (pk:publickey) (sk:secretkey) : option secretkey =\n let shared = Spec.Curve25519.scalarmult sk pk in\n if not (lbytes_eq shared (create 32 (u8 0))) then\n Some shared\n else None\n\n\nlet box_beforenm (pk:publickey) (sk:secretkey) : option key =\n let shared = ecdh pk sk in\n match shared with\n | Some shared -> Some (Spec.Salsa20.hsalsa20 shared (create 16 (u8 0)))\n | None -> None\n\n\nlet box_detached_afternm (k:key) (n:nonce) (m:bytes{length m / size_block <= max_size_t}) : tag & c:bytes{length c = length m} =\n secretbox_detached k n m\n\n\nlet box_detached (sk:secretkey) (pk:publickey) (n:nonce) (m:bytes{length m / size_block <= max_size_t}) : option (tag & c:bytes{length c = length m}) =\n let k = box_beforenm pk sk in\n match k with\n | Some k -> Some (box_detached_afternm k n m)\n | None -> None\n\n\nlet box_open_detached_afternm (k:key) (n:nonce) (t:tag) (c:bytes{length c / size_block <= max_size_t}) : option (m:bytes{length m = length c}) =\n secretbox_open_detached k n t c\n\n\nlet box_open_detached (pk:publickey) (sk:secretkey) (n:nonce) (t:tag) (c:bytes{length c / size_block <= max_size_t}) : option (m:bytes{length m = length c}) =\n let k = box_beforenm pk sk in\n match k with\n | Some k -> box_open_detached_afternm k n t c\n | None -> None\n\n\nlet box_easy_afternm (k:key) (n:nonce) (m:bytes{length m / size_block <= max_size_t}) : c:bytes{length c = size_tag + length m} =\n let (tg, c) = box_detached_afternm k n m in\n Seq.append tg c\n\nPartial implementation:\nlet box_easy\n (sk: secretkey)\n (pk: publickey)\n (n: nonce)\n (m: bytes{length m / size_block <= max_size_t})\n : option (c: bytes{length c = size_tag + length m}) =\n \n"
},
{
"role": "assistant",
"content": "let r = box_detached sk pk n m in\nmatch r with\n| Some (tg, c) -> Some (Seq.append tg c)\n| None -> None"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n va_b0: Vale.X64.Decls.va_code ->\n va_s0: Vale.X64.Decls.va_state ->\n alg: Vale.AES.AES_common_s.algorithm ->\n in_b: Vale.X64.Memory.buffer128 ->\n out_b: Vale.X64.Memory.buffer128 ->\n key: FStar.Seq.Base.seq Vale.X64.Memory.nat32 ->\n round_keys: FStar.Seq.Base.seq Vale.X64.Decls.quad32 ->\n keys_b: Vale.X64.Memory.buffer128 ->\n hkeys_b: Vale.X64.Memory.buffer128 ->\n h_LE: Vale.X64.Decls.quad32\n -> Prims.Ghost (Vale.X64.Decls.va_state * Vale.X64.Decls.va_fuel)\nEffect: Prims.Ghost\n\nContext:\nmodule Vale.AES.X64.GCMdecryptOpt\nopen Vale.Def.Prop_s\nopen Vale.Def.Opaque_s\nopen FStar.Seq\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Seq_s\nopen Vale.Def.Types_s\nopen Vale.Arch.Types\nopen Vale.Arch.HeapImpl\nopen Vale.AES.AES_s\nopen Vale.AES.GCTR_s\nopen Vale.AES.GCTR\nopen Vale.AES.GCM\nopen Vale.AES.GHash_s\nopen Vale.AES.GHash\nopen Vale.AES.GCM_s\nopen Vale.AES.X64.AES\nopen Vale.AES.GF128_s\nopen Vale.AES.GF128\nopen Vale.Poly1305.Math\nopen Vale.AES.GCM_helpers\nopen Vale.AES.X64.GHash\nopen Vale.AES.X64.GCTR\nopen Vale.X64.Machine_s\nopen Vale.X64.Memory\nopen Vale.X64.Stack_i\nopen Vale.X64.State\nopen Vale.X64.Decls\nopen Vale.X64.InsBasic\nopen Vale.X64.InsMem\nopen Vale.X64.InsVector\nopen Vale.X64.InsStack\nopen Vale.X64.InsAes\nopen Vale.X64.QuickCode\nopen Vale.X64.QuickCodes\nopen Vale.AES.X64.GF128_Mul\nopen Vale.X64.Stack\nopen Vale.X64.CPU_Features_s\nopen Vale.Math.Poly2.Bits_s\nopen Vale.AES.X64.AESopt\nopen Vale.AES.X64.AESGCM\nopen Vale.AES.X64.AESopt2\nopen Vale.Lib.Meta\nopen Vale.AES.X64.GCMencryptOpt\nopen Vale.AES.OptPublic\nopen Vale.Lib.Basic\n#reset-options \"--z3rlimit 20 --max_ifuel 0\"\n//-- Gcm_extra_bytes\n\nval va_code_Gcm_extra_bytes : alg:algorithm -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_extra_bytes alg =\n (va_Block (va_CCons (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0)\n (va_op_reg_opr64_reg64 rRax) 0 Secret) (va_CCons (va_code_Mov128 (va_op_xmm_xmm 10)\n (va_op_xmm_xmm 0)) (va_CCons (va_code_Ghash_extra_bytes ()) (va_CCons (va_code_Mov128\n (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_CCons (va_code_Pshufb (va_op_xmm_xmm 0)\n (va_op_xmm_xmm 9)) (va_CCons (va_code_AESEncryptBlock alg) (va_CCons (va_code_Pxor\n (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (va_CCons (va_code_Store128_buffer\n (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax) (va_op_xmm_xmm 10) 0 Secret)\n (va_CNil ()))))))))))\n\nval va_codegen_success_Gcm_extra_bytes : alg:algorithm -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_extra_bytes alg =\n (va_pbool_and (va_codegen_success_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0)\n (va_op_reg_opr64_reg64 rRax) 0 Secret) (va_pbool_and (va_codegen_success_Mov128 (va_op_xmm_xmm\n 10) (va_op_xmm_xmm 0)) (va_pbool_and (va_codegen_success_Ghash_extra_bytes ()) (va_pbool_and\n (va_codegen_success_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_pbool_and\n (va_codegen_success_Pshufb (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (va_pbool_and\n (va_codegen_success_AESEncryptBlock alg) (va_pbool_and (va_codegen_success_Pxor (va_op_xmm_xmm\n 10) (va_op_xmm_xmm 0)) (va_pbool_and (va_codegen_success_Store128_buffer\n (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax) (va_op_xmm_xmm 10) 0 Secret)\n (va_ttrue ())))))))))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_extra_bytes (va_mods:va_mods_t) (alg:algorithm) (inout_b:buffer128) (key:(seq\n nat32)) (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (total_bytes:nat)\n (old_hash:quad32) (completed_quads:(seq quad32)) (h_LE:quad32) : (va_quickCode unit\n (va_code_Gcm_extra_bytes alg)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in let (len:(va_int_range\n 1 1)) = 1 in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 188 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64\n rRax) 0 Secret inout_b 0) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 189 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov128 (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (fun (va_s:va_state) _ -> let\n (hash_input:quad32) = va_get_xmm 0 va_s in va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 193 column 22 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Ghash_extra_bytes hkeys_b total_bytes old_hash h_LE completed_quads) (fun\n (va_s:va_state) _ -> va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 194 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (FStar.Seq.Base.equal #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s)\n inout_b) (FStar.Seq.Base.create #quad32 1 hash_input)) (let (snap:(FStar.Seq.Base.seq\n Vale.X64.Decls.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s) inout_b in va_QSeq\n va_range1\n \"***** PRECONDITION NOT MET AT line 198 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 199 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Pshufb (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (fun (va_s:va_state) _ -> va_QBind\n va_range1\n \"***** PRECONDITION NOT MET AT line 200 column 20 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_AESEncryptBlock alg (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 11 va_s)) key\n round_keys keys_b) (fun (va_s:va_state) _ -> va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 201 column 26 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.AES_s.aes_encrypt_LE_reveal ()) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 204 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Pxor (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 205 column 20 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Store128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax)\n (va_op_xmm_xmm 10) 0 Secret inout_b 0) (fun (va_s:va_state) _ -> va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 207 column 24 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.gctr_partial_reveal ()) (va_QEmpty (()))))))))))))))\n\n\nval va_lemma_Gcm_extra_bytes : va_b0:va_code -> va_s0:va_state -> alg:algorithm ->\n inout_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128 -> total_bytes:nat -> old_hash:quad32 -> completed_quads:(seq quad32) ->\n h_LE:quad32\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_extra_bytes alg) va_s0 /\\ va_get_ok va_s0 /\\ (let\n (len:(va_int_range 1 1)) = 1 in sse_enabled /\\ Vale.X64.Decls.buffers_disjoint128 keys_b\n inout_b /\\ Vale.X64.Decls.buffers_disjoint128 hkeys_b inout_b /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) (va_get_reg64 rRax va_s0) inout_b\n len (va_get_mem_layout va_s0) Secret /\\ len == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 inout_b /\\ va_get_xmm 9 va_s0 == Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\ aes_reqs alg key round_keys\n keys_b (va_get_reg64 rR8 va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\\n pclmulqdq_enabled /\\ Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0\n va_s0) hkeys_b) (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret /\\ va_get_xmm 8 va_s0 == Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.AES.GHash.ghash_incremental0 h_LE old_hash completed_quads) /\\ FStar.Seq.Base.length\n #quad32 completed_quads == total_bytes `op_Division` 16 /\\ total_bytes < 16 `op_Multiply`\n FStar.Seq.Base.length #quad32 completed_quads + 16 /\\ va_get_reg64 rR10 va_s0 == total_bytes\n `op_Modulus` 16 /\\ total_bytes `op_Modulus` 16 =!= 0 /\\ (0 < total_bytes /\\ total_bytes < 16\n `op_Multiply` Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes) /\\ 16 `op_Multiply`\n (Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes - 1) < total_bytes)))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (let (len:(va_int_range 1 1)) = 1 in Vale.X64.Decls.modifies_buffer128 inout_b\n (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5 va_sM) /\\ Vale.AES.GCTR.gctr_partial alg len\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b) key (va_get_xmm 11 va_s0) /\\ (let raw_quads =\n FStar.Seq.Base.append #quad32 completed_quads (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0)\n inout_b) in let input_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads) 0 total_bytes in let padded_bytes =\n Vale.AES.GCTR_s.pad_to_128_bits input_bytes in let input_quads =\n Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in l_and (FStar.Seq.Base.length\n #Vale.Def.Types_s.quad32 input_quads > 0) (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8\n va_sM) == Vale.AES.GHash.ghash_incremental h_LE old_hash input_quads))) /\\ va_state_eq va_sM\n (va_update_flags va_sM (va_update_mem_heaplet 5 va_sM (va_update_xmm 10 va_sM (va_update_xmm 8\n va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5 va_sM (va_update_xmm 4\n va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1 va_sM (va_update_xmm 0\n va_sM (va_update_reg64 rR11 va_sM (va_update_reg64 rRcx va_sM (va_update_ok va_sM\n (va_update_mem va_sM va_s0))))))))))))))))))\n[@\"opaque_to_smt\"]\nlet va_lemma_Gcm_extra_bytes va_b0 va_s0 alg inout_b key round_keys keys_b hkeys_b total_bytes\n old_hash completed_quads h_LE =\n let (va_mods:va_mods_t) = [va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8;\n va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm\n 1; va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_ok; va_Mod_mem] in\n let va_qc = va_qcode_Gcm_extra_bytes va_mods alg inout_b key round_keys keys_b hkeys_b\n total_bytes old_hash completed_quads h_LE in\n let (va_sM, va_fM, va_g) = va_wp_sound_code_norm (va_code_Gcm_extra_bytes alg) va_qc va_s0 (fun\n va_s0 va_sM va_g -> let () = va_g in label va_range1\n \"***** POSTCONDITION NOT MET AT line 121 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\ (let (len:(va_int_range 1 1)) = 1 in label va_range1\n \"***** POSTCONDITION NOT MET AT line 174 column 55 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 177 column 95 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.AES.GCTR.gctr_partial alg len (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM) inout_b) key (va_get_xmm 11 va_s0)) /\\ label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 180 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let raw_quads = FStar.Seq.Base.append #quad32 completed_quads (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 181 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let input_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads) 0 total_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 182 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let padded_bytes = Vale.AES.GCTR_s.pad_to_128_bits input_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 183 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let input_quads = Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 186 column 59 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (l_and (FStar.Seq.Base.length #Vale.Def.Types_s.quad32 input_quads > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE old_hash input_quads)))))))) in\n assert_norm (va_qc.mods == va_mods);\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm\n 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1;\n va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_ok; va_Mod_mem]) va_sM va_s0;\n (va_sM, va_fM)\n\n[@ va_qattr]\nlet va_wp_Gcm_extra_bytes (alg:algorithm) (inout_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (total_bytes:nat) (old_hash:quad32)\n (completed_quads:(seq quad32)) (h_LE:quad32) (va_s0:va_state) (va_k:(va_state -> unit -> Type0))\n : Type0 =\n (va_get_ok va_s0 /\\ (let (len:(va_int_range 1 1)) = 1 in sse_enabled /\\\n Vale.X64.Decls.buffers_disjoint128 keys_b inout_b /\\ Vale.X64.Decls.buffers_disjoint128 hkeys_b\n inout_b /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) (va_get_reg64 rRax\n va_s0) inout_b len (va_get_mem_layout va_s0) Secret /\\ len == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 inout_b /\\ va_get_xmm 9 va_s0 == Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\ aes_reqs alg key round_keys\n keys_b (va_get_reg64 rR8 va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\\n pclmulqdq_enabled /\\ Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0\n va_s0) hkeys_b) (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret /\\ va_get_xmm 8 va_s0 == Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.AES.GHash.ghash_incremental0 h_LE old_hash completed_quads) /\\ FStar.Seq.Base.length\n #quad32 completed_quads == total_bytes `op_Division` 16 /\\ total_bytes < 16 `op_Multiply`\n FStar.Seq.Base.length #quad32 completed_quads + 16 /\\ va_get_reg64 rR10 va_s0 == total_bytes\n `op_Modulus` 16 /\\ total_bytes `op_Modulus` 16 =!= 0 /\\ (0 < total_bytes /\\ total_bytes < 16\n `op_Multiply` Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes) /\\ 16 `op_Multiply`\n (Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes - 1) < total_bytes) /\\ (forall\n (va_x_mem:vale_heap) (va_x_rcx:nat64) (va_x_r11:nat64) (va_x_xmm0:quad32) (va_x_xmm1:quad32)\n (va_x_xmm2:quad32) (va_x_xmm3:quad32) (va_x_xmm4:quad32) (va_x_xmm5:quad32) (va_x_xmm6:quad32)\n (va_x_xmm7:quad32) (va_x_xmm8:quad32) (va_x_xmm10:quad32) (va_x_heap5:vale_heap)\n (va_x_efl:Vale.X64.Flags.t) . let va_sM = va_upd_flags va_x_efl (va_upd_mem_heaplet 5\n va_x_heap5 (va_upd_xmm 10 va_x_xmm10 (va_upd_xmm 8 va_x_xmm8 (va_upd_xmm 7 va_x_xmm7\n (va_upd_xmm 6 va_x_xmm6 (va_upd_xmm 5 va_x_xmm5 (va_upd_xmm 4 va_x_xmm4 (va_upd_xmm 3 va_x_xmm3\n (va_upd_xmm 2 va_x_xmm2 (va_upd_xmm 1 va_x_xmm1 (va_upd_xmm 0 va_x_xmm0 (va_upd_reg64 rR11\n va_x_r11 (va_upd_reg64 rRcx va_x_rcx (va_upd_mem va_x_mem va_s0)))))))))))))) in va_get_ok\n va_sM /\\ (let (len:(va_int_range 1 1)) = 1 in Vale.X64.Decls.modifies_buffer128 inout_b\n (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5 va_sM) /\\ Vale.AES.GCTR.gctr_partial alg len\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b) key (va_get_xmm 11 va_s0) /\\ (let raw_quads =\n FStar.Seq.Base.append #quad32 completed_quads (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0)\n inout_b) in let input_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads) 0 total_bytes in let padded_bytes =\n Vale.AES.GCTR_s.pad_to_128_bits input_bytes in let input_quads =\n Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in l_and (FStar.Seq.Base.length\n #Vale.Def.Types_s.quad32 input_quads > 0) (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8\n va_sM) == Vale.AES.GHash.ghash_incremental h_LE old_hash input_quads))) ==> va_k va_sM (())))\n\n\nval va_wpProof_Gcm_extra_bytes : alg:algorithm -> inout_b:buffer128 -> key:(seq nat32) ->\n round_keys:(seq quad32) -> keys_b:buffer128 -> hkeys_b:buffer128 -> total_bytes:nat ->\n old_hash:quad32 -> completed_quads:(seq quad32) -> h_LE:quad32 -> va_s0:va_state ->\n va_k:(va_state -> unit -> Type0)\n -> Ghost (va_state & va_fuel & unit)\n (requires (va_t_require va_s0 /\\ va_wp_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b\n total_bytes old_hash completed_quads h_LE va_s0 va_k))\n (ensures (fun (va_sM, va_f0, va_g) -> va_t_ensure (va_code_Gcm_extra_bytes alg) ([va_Mod_flags;\n va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5;\n va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR11;\n va_Mod_reg64 rRcx; va_Mod_mem]) va_s0 va_k ((va_sM, va_f0, va_g))))\n[@\"opaque_to_smt\"]\nlet va_wpProof_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b total_bytes old_hash\n completed_quads h_LE va_s0 va_k =\n let (va_sM, va_f0) = va_lemma_Gcm_extra_bytes (va_code_Gcm_extra_bytes alg) va_s0 alg inout_b key\n round_keys keys_b hkeys_b total_bytes old_hash completed_quads h_LE in\n va_lemma_upd_update va_sM;\n assert (va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 5 va_sM (va_update_xmm 10\n va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5\n va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1\n va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR11 va_sM (va_update_reg64 rRcx va_sM\n (va_update_ok va_sM (va_update_mem va_sM va_s0)))))))))))))))));\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm\n 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1;\n va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_mem]) va_sM va_s0;\n let va_g = () in\n (va_sM, va_f0, va_g)\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_quick_Gcm_extra_bytes (alg:algorithm) (inout_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (total_bytes:nat) (old_hash:quad32)\n (completed_quads:(seq quad32)) (h_LE:quad32) : (va_quickCode unit (va_code_Gcm_extra_bytes alg)) =\n (va_QProc (va_code_Gcm_extra_bytes alg) ([va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10;\n va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm\n 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_mem])\n (va_wp_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b total_bytes old_hash\n completed_quads h_LE) (va_wpProof_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b\n total_bytes old_hash completed_quads h_LE))\n//--\n//-- Gcm_blocks128\n\nval va_code_Gcm_blocks128 : alg:algorithm -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_blocks128 alg =\n (va_Block (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRbx) (va_op_opr64_reg64 rRdi))\n (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR12) (va_op_opr64_reg64 rRdx)) (va_CCons\n (va_code_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRax)) (va_CCons\n (va_code_Ghash_buffer ()) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRdi)\n (va_op_opr64_reg64 rRbx)) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRdx)\n (va_op_opr64_reg64 rR12)) (va_CCons (va_code_Gctr_blocks128 alg) (va_CNil ())))))))))\n\nval va_codegen_success_Gcm_blocks128 : alg:algorithm -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_blocks128 alg =\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rRbx) (va_op_opr64_reg64 rRdi))\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR12) (va_op_opr64_reg64 rRdx))\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRax))\n (va_pbool_and (va_codegen_success_Ghash_buffer ()) (va_pbool_and (va_codegen_success_Mov64\n (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRbx)) (va_pbool_and (va_codegen_success_Mov64\n (va_op_dst_opr64_reg64 rRdx) (va_op_opr64_reg64 rR12)) (va_pbool_and\n (va_codegen_success_Gctr_blocks128 alg) (va_ttrue ()))))))))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_blocks128 (va_mods:va_mods_t) (alg:algorithm) (in_b:buffer128) (out_b:buffer128)\n (key:(seq nat32)) (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (h_LE:quad32)\n : (va_quickCode unit (va_code_Gcm_blocks128 alg)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 274 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRbx) (va_op_opr64_reg64 rRdi)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 275 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR12) (va_op_opr64_reg64 rRdx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 276 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRax)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 277 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Ghash_buffer hkeys_b in_b h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8\n va_old_s))) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 278 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRbx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 279 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRdx) (va_op_opr64_reg64 rR12)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 280 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gctr_blocks128 alg in_b out_b key round_keys keys_b) (va_QEmpty (()))))))))))\n\n\nval va_lemma_Gcm_blocks128 : va_b0:va_code -> va_s0:va_state -> alg:algorithm -> in_b:buffer128 ->\n out_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128 -> h_LE:quad32\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_blocks128 alg) va_s0 /\\ va_get_ok va_s0 /\\\n (sse_enabled /\\ Vale.X64.Decls.buffers_disjoint128 keys_b out_b /\\\n Vale.X64.Decls.buffers_disjoint128 hkeys_b out_b /\\ (Vale.X64.Decls.buffers_disjoint128 in_b\n out_b \\/ in_b == out_b) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0)\n (va_get_reg64 rRax va_s0) in_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1 va_s0) (va_get_reg64 rRdi va_s0) out_b\n (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\ va_get_reg64 rRax va_s0 + 16\n `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply`\n va_get_reg64 rRdx va_s0 < pow2_64 /\\ l_and (Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 in_b == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out_b)\n (Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b < pow2_32) /\\ va_get_reg64 rRdx\n va_s0 == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b /\\ va_get_xmm 9 va_s0 ==\n Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\\n va_get_reg64 rRdx va_s0 < pow2_32 /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rR8\n va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret)))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (Vale.X64.Decls.modifies_buffer128 out_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM) /\\ Vale.AES.GCTR.gctr_partial alg (va_get_reg64 rRdx va_s0) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b) key\n (va_get_xmm 11 va_s0) /\\ va_get_xmm 11 va_sM == Vale.AES.GCTR.inc32lite (va_get_xmm 11 va_s0)\n (va_get_reg64 rRdx va_s0) /\\ (va_get_reg64 rRdx va_s0 == 0 ==> l_and (va_get_xmm 8 va_sM ==\n va_get_xmm 8 va_s0) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b ==\n Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) out_b)) /\\ (va_get_reg64 rRdx va_s0 > 0 ==>\n l_and (va_get_reg64 rRdx va_s0 <= FStar.Seq.Base.length #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) ==> FStar.Seq.Base.length\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.slice #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) 0 (va_get_reg64 rRdx va_s0)) > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s0)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b)))) /\\ va_state_eq va_sM (va_update_flags va_sM\n (va_update_mem_heaplet 1 va_sM (va_update_xmm 10 va_sM (va_update_xmm 11 va_sM (va_update_xmm 8\n va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5 va_sM (va_update_xmm 4\n va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1 va_sM (va_update_xmm 0\n va_sM (va_update_reg64 rR12 va_sM (va_update_reg64 rRdx va_sM (va_update_reg64 rR10 va_sM\n (va_update_reg64 rR11 va_sM (va_update_reg64 rRdi va_sM (va_update_reg64 rRbx va_sM\n (va_update_ok va_sM (va_update_mem va_sM va_s0)))))))))))))))))))))))\n\nPartial implementation:\nlet va_lemma_Gcm_blocks128 va_b0 va_s0 alg in_b out_b key round_keys keys_b hkeys_b h_LE =\n \n"
},
{
"role": "assistant",
"content": "let va_mods:va_mods_t =\n [\n va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0;\n va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10; va_Mod_reg64 rR11; va_Mod_reg64 rRdi;\n va_Mod_reg64 rRbx; va_Mod_ok; va_Mod_mem\n ]\nin\nlet va_qc = va_qcode_Gcm_blocks128 va_mods alg in_b out_b key round_keys keys_b hkeys_b h_LE in\nlet va_sM, va_fM, va_g =\n va_wp_sound_code_norm (va_code_Gcm_blocks128 alg)\n va_qc\n va_s0\n (fun va_s0 va_sM va_g ->\n let () = va_g in\n label va_range1\n \"***** POSTCONDITION NOT MET AT line 210 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\\n (label va_range1\n \"***** POSTCONDITION NOT MET AT line 255 column 53 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out_b\n (va_get_mem_heaplet 1 va_s0)\n (va_get_mem_heaplet 1 va_sM)) /\\\n label va_range1\n \"***** POSTCONDITION NOT MET AT line 261 column 95 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.AES.GCTR.gctr_partial alg\n (va_get_reg64 rRdx va_s0)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b)\n key\n (va_get_xmm 11 va_s0)) /\\\n label va_range1\n \"***** POSTCONDITION NOT MET AT line 262 column 45 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_xmm 11 va_sM ==\n Vale.AES.GCTR.inc32lite (va_get_xmm 11 va_s0) (va_get_reg64 rRdx va_s0)) /\\\n label va_range1\n \"***** POSTCONDITION NOT MET AT line 265 column 93 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRdx va_s0 == 0 ==>\n l_and (va_get_xmm 8 va_sM == va_get_xmm 8 va_s0)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b ==\n Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) out_b)) /\\\n label va_range1\n \"***** POSTCONDITION NOT MET AT line 267 column 131 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRdx va_s0 > 0 ==>\n l_and (va_get_reg64 rRdx va_s0 <=\n FStar.Seq.Base.length #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) ==>\n FStar.Seq.Base.length #Vale.X64.Decls.quad32\n (FStar.Seq.Base.slice #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b)\n 0\n (va_get_reg64 rRdx va_s0)) >\n 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) ==\n Vale.AES.GHash.ghash_incremental h_LE\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s0))\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b)))))\nin\nassert_norm (va_qc.mods == va_mods);\nva_lemma_norm_mods ([\n va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1;\n va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10; va_Mod_reg64 rR11;\n va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_ok; va_Mod_mem\n ])\n va_sM\n va_s0;\n(va_sM, va_fM)"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n h0: FStar.Monotonic.HyperStack.mem ->\n h1: FStar.Monotonic.HyperStack.mem ->\n ll0: LowStar.Lib.LinkedList.t a ->\n ll1: LowStar.Lib.LinkedList.t a ->\n l0: Prims.list a ->\n l1: Prims.list a\n -> FStar.Pervasives.Lemma\n (requires\n LowStar.Lib.LinkedList.well_formed h0 ll0 l0 /\\ LowStar.Lib.LinkedList.well_formed h1 ll1 l1 /\\\n LowStar.Lib.LinkedList.cells h0 ll0 l0 == LowStar.Lib.LinkedList.cells h1 ll1 l1)\n (ensures ll0 == ll1)\n [\n SMTPat (LowStar.Lib.LinkedList.cells h0 ll0 l0);\n SMTPat (LowStar.Lib.LinkedList.cells h1 ll1 l1)\n ]\nEffect: FStar.Pervasives.Lemma\n\nContext:\nmodule LowStar.Lib.LinkedList\n\n/// This module, oftentimes referred to as LL1 in other parts of the\n/// documentation, provides a low-level view over a linked list. Unless you\n/// intend on modifying the structure of the linked list (e.g. removing cells)\n/// or iterating over each cell, your are probably better off using LL2.\n///\n/// This module is intentionally not relying on a tight abstraction, to i) permit\n/// direct-style iteration by clients and ii) allow stating some SMTPat lemmas\n/// only on LL1 rather than replicating them in LL2 (going from LL2 to LL1 can\n/// happen automatically at the SMT level since LL2 does not have an fsti either).\n///\n/// Some open questions relative to the design of this module:\n/// - should ``v`` be defined directly in terms of ``cells`` rather than have a\n/// recursive definition along with an equivalence lemma?\n/// - should ``footprint`` be defined directly in terms of ``cells`` rather than\n/// have a recursive definition along with an equivalence lemma?\n/// - is it worth the trouble defining a footprint? my intuition is that it's\n/// fine and allows some clients to do very precise reasoning -- other modules\n/// (e.g. AssocList), when establishing abstraction boundaries, can simplify\n/// things and get rid of the precise footprint reasoning in favor of a static\n/// region-based footprint\n///\n/// Some remaining work items:\n/// - switch to C.Loops.while\n\nopen LowStar.BufferOps\n\nmodule B = LowStar.Buffer\nmodule HS = FStar.HyperStack\nmodule G = FStar.Ghost\nmodule L = FStar.List.Tot\nmodule U32 = FStar.UInt32\nmodule ST = FStar.HyperStack.ST\n\nopen FStar.HyperStack.ST\n\n/// Breaking from my own recommendation here... using fuel 1 / ifuel 1 since\n/// everything does recursive list reasoning.\n#set-options \"--__no_positivity --fuel 1 --ifuel 1\"\n\n/// Definition of the Low* type\n/// ---------------------------\nnoeq\ntype t (a: Type0) =\n B.pointer_or_null (cell a)\n\nand cell (a: Type0) = {\n next: t a;\n data: a;\n}\n\n/// Unlike the canonical style that associates a ``v`` function to any stateful\n/// representation, we use a relation here. Writing a function is just\n/// impossible, since we can't prove termination of ``v`` owing to the fact that\n/// there may be an infinite number of cells in the heap.\n///\n/// LL2 is the module that switches to the function style.\n///\n/// Note: no need to use erased here for ``l`` since the function is already in ``GTot``.\nlet rec well_formed #a (h: HS.mem) (c: t a) (l: list a): GTot Type0 (decreases l) =\n B.live h c /\\ (\n match l with\n | [] ->\n B.g_is_null c\n | a :: q ->\n B.length c == 1 /\\ (\n let { data=data; next=next } = B.get h c 0 in\n a == data /\\\n well_formed h next (G.hide q)\n ))\n\n/// Precise reasoning via the cells predicate\n/// -----------------------------------------\n///\n/// The modifies clauses are given either on the footprint of the list, or on a\n/// region that includes the footprint of the list (higher up the stack, e.g.\n/// LL2). This is good, but does not rule out \"stupid\" implementations that\n/// might re-allocate every single list cell when, say, doing a pop.\n///\n/// For advanced usages which take pointers directly to individual list cells,\n/// it's important to rule out these cases. We thus define the ``cells``\n/// predicate for that purpose.\n\nlet rec cells #a (h: HS.mem) (c: t a) (l: list a): Ghost (list (B.pointer (cell a)))\n (requires well_formed h c l)\n (ensures fun _ -> True)\n (decreases l)\n=\n if B.g_is_null c then\n []\n else\n c :: cells h (B.deref h c).next (List.Tot.tl l)\n\n/// A missing bit of automated reasoning\n/// ------------------------------------\n\n/// This should be absolutely trivial, yet, because there's no pattern on\n/// null_unique, we have to do the case analysis and the decomposition\n/// ourselves. This is quite tedious, as it usually takes a while (at least it\n/// did for me!) to realize that this property does not hold automatically.\n///\n/// I'm setting an SMTPat on this lemma because it's absolutely impossible to\n\nPartial implementation:\nlet same_cells_same_pointer\n #a\n (h0: HS.mem)\n (h1: HS.mem)\n (ll0: t a)\n (ll1: t a)\n (l0: list a)\n (l1: list a)\n : Lemma\n (requires\n (well_formed h0 ll0 l0 /\\ well_formed h1 ll1 l1 /\\ cells h0 ll0 l0 == cells h1 ll1 l1))\n (ensures (ll0 == ll1))\n [SMTPat (cells h0 ll0 l0); SMTPat (cells h1 ll1 l1)] =\n \n"
},
{
"role": "assistant",
"content": "if B.g_is_null ll0 && B.g_is_null ll1\nthen\n (B.null_unique ll0;\n B.null_unique ll1)\nelse if not (B.g_is_null ll0) && not (B.g_is_null ll1) then () else false_elim ()"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: f: (_: b -> _: a -> Prims.GTot b) -> xs: Prims.list a -> acc: b -> Prims.Ghost b\nEffect: Prims.Ghost\n\nContext:\nmodule LowStar.Lib.LinkedList\n\n/// This module, oftentimes referred to as LL1 in other parts of the\n/// documentation, provides a low-level view over a linked list. Unless you\n/// intend on modifying the structure of the linked list (e.g. removing cells)\n/// or iterating over each cell, your are probably better off using LL2.\n///\n/// This module is intentionally not relying on a tight abstraction, to i) permit\n/// direct-style iteration by clients and ii) allow stating some SMTPat lemmas\n/// only on LL1 rather than replicating them in LL2 (going from LL2 to LL1 can\n/// happen automatically at the SMT level since LL2 does not have an fsti either).\n///\n/// Some open questions relative to the design of this module:\n/// - should ``v`` be defined directly in terms of ``cells`` rather than have a\n/// recursive definition along with an equivalence lemma?\n/// - should ``footprint`` be defined directly in terms of ``cells`` rather than\n/// have a recursive definition along with an equivalence lemma?\n/// - is it worth the trouble defining a footprint? my intuition is that it's\n/// fine and allows some clients to do very precise reasoning -- other modules\n/// (e.g. AssocList), when establishing abstraction boundaries, can simplify\n/// things and get rid of the precise footprint reasoning in favor of a static\n/// region-based footprint\n///\n/// Some remaining work items:\n/// - switch to C.Loops.while\n\nopen LowStar.BufferOps\n\nmodule B = LowStar.Buffer\nmodule HS = FStar.HyperStack\nmodule G = FStar.Ghost\nmodule L = FStar.List.Tot\nmodule U32 = FStar.UInt32\nmodule ST = FStar.HyperStack.ST\n\nopen FStar.HyperStack.ST\n\n/// Breaking from my own recommendation here... using fuel 1 / ifuel 1 since\n/// everything does recursive list reasoning.\n#set-options \"--__no_positivity --fuel 1 --ifuel 1\"\n\n/// Definition of the Low* type\n/// ---------------------------\nnoeq\ntype t (a: Type0) =\n B.pointer_or_null (cell a)\n\nand cell (a: Type0) = {\n next: t a;\n data: a;\n}\n\n/// Unlike the canonical style that associates a ``v`` function to any stateful\n/// representation, we use a relation here. Writing a function is just\n/// impossible, since we can't prove termination of ``v`` owing to the fact that\n/// there may be an infinite number of cells in the heap.\n///\n/// LL2 is the module that switches to the function style.\n///\n/// Note: no need to use erased here for ``l`` since the function is already in ``GTot``.\nlet rec well_formed #a (h: HS.mem) (c: t a) (l: list a): GTot Type0 (decreases l) =\n B.live h c /\\ (\n match l with\n | [] ->\n B.g_is_null c\n | a :: q ->\n B.length c == 1 /\\ (\n let { data=data; next=next } = B.get h c 0 in\n a == data /\\\n well_formed h next (G.hide q)\n ))\n\n/// Precise reasoning via the cells predicate\n/// -----------------------------------------\n///\n/// The modifies clauses are given either on the footprint of the list, or on a\n/// region that includes the footprint of the list (higher up the stack, e.g.\n/// LL2). This is good, but does not rule out \"stupid\" implementations that\n/// might re-allocate every single list cell when, say, doing a pop.\n///\n/// For advanced usages which take pointers directly to individual list cells,\n/// it's important to rule out these cases. We thus define the ``cells``\n/// predicate for that purpose.\n\nlet rec cells #a (h: HS.mem) (c: t a) (l: list a): Ghost (list (B.pointer (cell a)))\n (requires well_formed h c l)\n (ensures fun _ -> True)\n (decreases l)\n=\n if B.g_is_null c then\n []\n else\n c :: cells h (B.deref h c).next (List.Tot.tl l)\n\n/// A missing bit of automated reasoning\n/// ------------------------------------\n\n/// This should be absolutely trivial, yet, because there's no pattern on\n/// null_unique, we have to do the case analysis and the decomposition\n/// ourselves. This is quite tedious, as it usually takes a while (at least it\n/// did for me!) to realize that this property does not hold automatically.\n///\n/// I'm setting an SMTPat on this lemma because it's absolutely impossible to\n/// get any work done without it.\nlet same_cells_same_pointer #a (h0 h1: HS.mem) (ll0 ll1: t a) (l0 l1: list a): Lemma\n (requires (\n well_formed h0 ll0 l0 /\\\n well_formed h1 ll1 l1 /\\\n cells h0 ll0 l0 == cells h1 ll1 l1))\n (ensures (\n ll0 == ll1))\n [ SMTPat (cells h0 ll0 l0); SMTPat (cells h1 ll1 l1) ]\n=\n if B.g_is_null ll0 && B.g_is_null ll1 then begin\n B.null_unique ll0;\n B.null_unique ll1\n end else if not (B.g_is_null ll0) && not (B.g_is_null ll1) then begin\n ()\n end else\n false_elim ()\n\n/// Classic stateful reasoning lemmas\n/// ---------------------------------\n///\n/// This essentially follows ``stateful``, the type class of a mutable piece of\n/// state defined using the modifies-theory of Low*v2 and equipped with the\n/// right lemmas to enable usage by clients. See\n/// hacl-star/code/streaming/Hacl.Streaming.Interface.fst for more context.\n\nval footprint: (#a: Type) -> (h: HS.mem) -> (l: t a) -> (n: list a) -> Ghost B.loc\n (requires (well_formed h l n))\n (ensures (fun refs -> True))\n (decreases n)\n\nlet rec footprint #a h l n =\n if B.g_is_null l\n then B.loc_none\n else\n let {next = next} = B.get h l 0 in\n let refs = footprint h next (G.hide (L.tl n)) in\n B.loc_union (B.loc_addr_of_buffer l) refs\n\n/// Departing from LinkedList4 here. I prefer to bolt these into the invariant\n/// rather than requiring on some wizard lemmas from LowStar.Monotonic.Buffer to\n/// deduce these.\nval cells_pairwise_disjoint: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_pairwise_disjoint #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.loc_disjoint (B.loc_addr_of_buffer l) (footprint h next (L.tl n)) /\\\n cells_pairwise_disjoint h next (L.tl n)\n\nval cells_live_freeable: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_live_freeable #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.live h l /\\ B.freeable l /\\\n cells_live_freeable h next (L.tl n)\n\nval invariant: #a:Type -> h:HS.mem -> l: t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\nlet invariant #a h l n =\n cells_live_freeable h l n /\\ cells_pairwise_disjoint h l n\n\n/// Normally this would be automatic and writing a custom lemma for that stuff\n/// would be needed only when writing an fsti, however, as stated earlier,\n/// recursion makes everything more difficult and this absolutely needs to be in\n/// scope otherwise the modifies-clause theory cannot deduce disjointness of\n/// fresh allocations w.r.t. the footprint.\nlet rec invariant_loc_in_footprint (#a: Type) (h: HS.mem) (l: t a) (n: list a): Lemma\n (requires well_formed h l n)\n (ensures B.loc_in (footprint h l n) h)\n (decreases n)\n [ SMTPat (invariant h l n) ]\n=\n if B.g_is_null l then\n ()\n else\n let next = (B.deref h l).next in\n invariant_loc_in_footprint h next (L.tl n)\n\n/// Another absolutely essential lemma, for which interestingly the pattern\n/// \"invariant h1 l n\" does not work, but well_formed does... I guess that's\n/// alright.\nlet rec frame (#a: Type) (l: t a) (n: list a) (r: B.loc) (h0 h1: HS.mem): Lemma\n (requires (\n well_formed h0 l n /\\\n invariant h0 l n /\\\n B.loc_disjoint r (footprint h0 l n) /\\\n B.modifies r h0 h1\n ))\n (ensures (\n well_formed h1 l n /\\\n footprint h1 l n == footprint h0 l n /\\\n cells h1 l n == cells h0 l n /\\\n invariant h1 l n\n ))\n (decreases n)\n [ SMTPat (well_formed h1 l n); SMTPat (B.modifies r h0 h1) ]\n=\n if B.g_is_null l then\n ()\n else\n frame (B.deref h0 l).next (L.tl n) r h0 h1\n\n/// Lemmas for working with linked lists\n/// ------------------------------------\n///\n/// Since it's a recursive data structure, a lot of the automated reasoning from\n/// LowStar.Buffer doesn't work since it requires inductions. Some helpful\n/// properties are therefore proved here.\n\nlet rec length_functional #a (h: HS.mem) (c: t a) (l1 l2: list a):\n Lemma\n (requires (well_formed h c l1 /\\ well_formed h c l2))\n (ensures (l1 == l2))\n (decreases (G.reveal l1))\n [ SMTPat (well_formed h c l1); SMTPat (well_formed h c l2) ] =\n if B.g_is_null c\n then ()\n else\n let { next=next } = B.get h c 0 in\n length_functional h next (G.hide (L.tl l1)) (G.hide (L.tl l2))\n\n/// The footprint is based on loc_addr_of_buffer so that we can write a free\n/// operation that operates on the footprint. However, this invalidates most of\n/// the helper lemmas for disjointness that were present in the previous\n/// iteration of this module, named LinkedList4.fst. It's ok, we bake\n/// disjointness directly onto the invariant and it works just as well.\n\n/// Connection between ``cells`` and ``v``\n/// --------------------------------------\n\n/// Redefining some helpers.\n#push-options \"--ifuel 1\"\nlet rec gmap #a #b (f: a -> GTot b) (xs: list a): GTot (list b) =\n match xs with\n | [] -> []\n | x :: xs -> f x :: gmap f xs\n\n/// \"As we all know\", right folds are the easiest to work with because they\n\nPartial implementation:\nlet rec gfold_right #a #b (f: (b -> a -> GTot b)) (xs: list a) (acc: b)\n : Ghost b (requires True) (ensures fun _ -> True) (decreases xs) =\n \n"
},
{
"role": "assistant",
"content": "match xs with\n| [] -> acc\n| x :: xs -> f (gfold_right f xs acc) x"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n pk: Spec.Box.secretkey ->\n sk: Spec.Box.publickey ->\n n: Spec.SecretBox.nonce ->\n c:\n Lib.ByteSequence.bytes\n { Lib.Sequence.length c >= Spec.SecretBox.size_tag /\\\n (Lib.Sequence.length c - Spec.SecretBox.size_tag) / Spec.SecretBox.size_block <=\n Lib.IntTypes.max_size_t }\n -> FStar.Pervasives.Native.option (m:\n Lib.ByteSequence.bytes\n {Lib.Sequence.length m = Lib.Sequence.length c - Spec.SecretBox.size_tag})\nEffect: Prims.Tot\n\nContext:\nmodule Spec.Box\n\nopen FStar.Mul\nopen Lib.IntTypes\nopen Lib.Sequence\nopen Lib.ByteSequence\n\nopen Spec.SecretBox\n\n#set-options \"--z3rlimit 20 --max_fuel 0 --max_ifuel 0\"\n\n(* Constants *)\nlet size_publickey = 32 (* in bytes *)\nlet size_secretkey = 32 (* in bytes *)\n\ntype publickey = lbytes size_publickey\ntype secretkey = lbytes size_secretkey\n\n\nlet ecdh (pk:publickey) (sk:secretkey) : option secretkey =\n let shared = Spec.Curve25519.scalarmult sk pk in\n if not (lbytes_eq shared (create 32 (u8 0))) then\n Some shared\n else None\n\n\nlet box_beforenm (pk:publickey) (sk:secretkey) : option key =\n let shared = ecdh pk sk in\n match shared with\n | Some shared -> Some (Spec.Salsa20.hsalsa20 shared (create 16 (u8 0)))\n | None -> None\n\n\nlet box_detached_afternm (k:key) (n:nonce) (m:bytes{length m / size_block <= max_size_t}) : tag & c:bytes{length c = length m} =\n secretbox_detached k n m\n\n\nlet box_detached (sk:secretkey) (pk:publickey) (n:nonce) (m:bytes{length m / size_block <= max_size_t}) : option (tag & c:bytes{length c = length m}) =\n let k = box_beforenm pk sk in\n match k with\n | Some k -> Some (box_detached_afternm k n m)\n | None -> None\n\n\nlet box_open_detached_afternm (k:key) (n:nonce) (t:tag) (c:bytes{length c / size_block <= max_size_t}) : option (m:bytes{length m = length c}) =\n secretbox_open_detached k n t c\n\n\nlet box_open_detached (pk:publickey) (sk:secretkey) (n:nonce) (t:tag) (c:bytes{length c / size_block <= max_size_t}) : option (m:bytes{length m = length c}) =\n let k = box_beforenm pk sk in\n match k with\n | Some k -> box_open_detached_afternm k n t c\n | None -> None\n\n\nlet box_easy_afternm (k:key) (n:nonce) (m:bytes{length m / size_block <= max_size_t}) : c:bytes{length c = size_tag + length m} =\n let (tg, c) = box_detached_afternm k n m in\n Seq.append tg c\n\n\nlet box_easy (sk:secretkey) (pk:publickey) (n:nonce) (m:bytes{length m / size_block <= max_size_t}) : option (c:bytes{length c = size_tag + length m}) =\n let r = box_detached sk pk n m in\n match r with\n | Some (tg, c) -> Some (Seq.append tg c)\n | None -> None\n\n\nlet box_open_easy_afternm (k:key) (n:nonce) (c:bytes{length c >= size_tag /\\ (length c - size_tag) / size_block <= max_size_t}) : option (m:bytes{length m = length c - size_tag}) =\n let tg = Seq.slice c 0 size_tag in\n let e = Seq.slice c size_tag (length c) in\n box_open_detached_afternm k n tg e\n\nPartial implementation:\nlet box_open_easy\n (pk: secretkey)\n (sk: publickey)\n (n: nonce)\n (c: bytes{length c >= size_tag /\\ (length c - size_tag) / size_block <= max_size_t})\n : option (m: bytes{length m = length c - size_tag}) =\n \n"
},
{
"role": "assistant",
"content": "let tg = Seq.slice c 0 size_tag in\nlet e = Seq.slice c size_tag (length c) in\nbox_open_detached pk sk n tg e"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: h: FStar.Monotonic.HyperStack.mem -> ll: LowStar.Lib.LinkedList.t a -> l: Prims.list a\n -> FStar.Pervasives.Lemma (requires LowStar.Lib.LinkedList.well_formed h ll l)\n (ensures\n LowStar.Lib.LinkedList.footprint h ll l ==\n LowStar.Lib.LinkedList.gfold_right LowStar.Monotonic.Buffer.loc_union\n (LowStar.Lib.LinkedList.gmap LowStar.Monotonic.Buffer.loc_addr_of_buffer\n (LowStar.Lib.LinkedList.cells h ll l))\n LowStar.Monotonic.Buffer.loc_none)\n (decreases l)\nEffect: FStar.Pervasives.Lemma\n\nContext:\nmodule LowStar.Lib.LinkedList\n\n/// This module, oftentimes referred to as LL1 in other parts of the\n/// documentation, provides a low-level view over a linked list. Unless you\n/// intend on modifying the structure of the linked list (e.g. removing cells)\n/// or iterating over each cell, your are probably better off using LL2.\n///\n/// This module is intentionally not relying on a tight abstraction, to i) permit\n/// direct-style iteration by clients and ii) allow stating some SMTPat lemmas\n/// only on LL1 rather than replicating them in LL2 (going from LL2 to LL1 can\n/// happen automatically at the SMT level since LL2 does not have an fsti either).\n///\n/// Some open questions relative to the design of this module:\n/// - should ``v`` be defined directly in terms of ``cells`` rather than have a\n/// recursive definition along with an equivalence lemma?\n/// - should ``footprint`` be defined directly in terms of ``cells`` rather than\n/// have a recursive definition along with an equivalence lemma?\n/// - is it worth the trouble defining a footprint? my intuition is that it's\n/// fine and allows some clients to do very precise reasoning -- other modules\n/// (e.g. AssocList), when establishing abstraction boundaries, can simplify\n/// things and get rid of the precise footprint reasoning in favor of a static\n/// region-based footprint\n///\n/// Some remaining work items:\n/// - switch to C.Loops.while\n\nopen LowStar.BufferOps\n\nmodule B = LowStar.Buffer\nmodule HS = FStar.HyperStack\nmodule G = FStar.Ghost\nmodule L = FStar.List.Tot\nmodule U32 = FStar.UInt32\nmodule ST = FStar.HyperStack.ST\n\nopen FStar.HyperStack.ST\n\n/// Breaking from my own recommendation here... using fuel 1 / ifuel 1 since\n/// everything does recursive list reasoning.\n#set-options \"--__no_positivity --fuel 1 --ifuel 1\"\n\n/// Definition of the Low* type\n/// ---------------------------\nnoeq\ntype t (a: Type0) =\n B.pointer_or_null (cell a)\n\nand cell (a: Type0) = {\n next: t a;\n data: a;\n}\n\n/// Unlike the canonical style that associates a ``v`` function to any stateful\n/// representation, we use a relation here. Writing a function is just\n/// impossible, since we can't prove termination of ``v`` owing to the fact that\n/// there may be an infinite number of cells in the heap.\n///\n/// LL2 is the module that switches to the function style.\n///\n/// Note: no need to use erased here for ``l`` since the function is already in ``GTot``.\nlet rec well_formed #a (h: HS.mem) (c: t a) (l: list a): GTot Type0 (decreases l) =\n B.live h c /\\ (\n match l with\n | [] ->\n B.g_is_null c\n | a :: q ->\n B.length c == 1 /\\ (\n let { data=data; next=next } = B.get h c 0 in\n a == data /\\\n well_formed h next (G.hide q)\n ))\n\n/// Precise reasoning via the cells predicate\n/// -----------------------------------------\n///\n/// The modifies clauses are given either on the footprint of the list, or on a\n/// region that includes the footprint of the list (higher up the stack, e.g.\n/// LL2). This is good, but does not rule out \"stupid\" implementations that\n/// might re-allocate every single list cell when, say, doing a pop.\n///\n/// For advanced usages which take pointers directly to individual list cells,\n/// it's important to rule out these cases. We thus define the ``cells``\n/// predicate for that purpose.\n\nlet rec cells #a (h: HS.mem) (c: t a) (l: list a): Ghost (list (B.pointer (cell a)))\n (requires well_formed h c l)\n (ensures fun _ -> True)\n (decreases l)\n=\n if B.g_is_null c then\n []\n else\n c :: cells h (B.deref h c).next (List.Tot.tl l)\n\n/// A missing bit of automated reasoning\n/// ------------------------------------\n\n/// This should be absolutely trivial, yet, because there's no pattern on\n/// null_unique, we have to do the case analysis and the decomposition\n/// ourselves. This is quite tedious, as it usually takes a while (at least it\n/// did for me!) to realize that this property does not hold automatically.\n///\n/// I'm setting an SMTPat on this lemma because it's absolutely impossible to\n/// get any work done without it.\nlet same_cells_same_pointer #a (h0 h1: HS.mem) (ll0 ll1: t a) (l0 l1: list a): Lemma\n (requires (\n well_formed h0 ll0 l0 /\\\n well_formed h1 ll1 l1 /\\\n cells h0 ll0 l0 == cells h1 ll1 l1))\n (ensures (\n ll0 == ll1))\n [ SMTPat (cells h0 ll0 l0); SMTPat (cells h1 ll1 l1) ]\n=\n if B.g_is_null ll0 && B.g_is_null ll1 then begin\n B.null_unique ll0;\n B.null_unique ll1\n end else if not (B.g_is_null ll0) && not (B.g_is_null ll1) then begin\n ()\n end else\n false_elim ()\n\n/// Classic stateful reasoning lemmas\n/// ---------------------------------\n///\n/// This essentially follows ``stateful``, the type class of a mutable piece of\n/// state defined using the modifies-theory of Low*v2 and equipped with the\n/// right lemmas to enable usage by clients. See\n/// hacl-star/code/streaming/Hacl.Streaming.Interface.fst for more context.\n\nval footprint: (#a: Type) -> (h: HS.mem) -> (l: t a) -> (n: list a) -> Ghost B.loc\n (requires (well_formed h l n))\n (ensures (fun refs -> True))\n (decreases n)\n\nlet rec footprint #a h l n =\n if B.g_is_null l\n then B.loc_none\n else\n let {next = next} = B.get h l 0 in\n let refs = footprint h next (G.hide (L.tl n)) in\n B.loc_union (B.loc_addr_of_buffer l) refs\n\n/// Departing from LinkedList4 here. I prefer to bolt these into the invariant\n/// rather than requiring on some wizard lemmas from LowStar.Monotonic.Buffer to\n/// deduce these.\nval cells_pairwise_disjoint: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_pairwise_disjoint #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.loc_disjoint (B.loc_addr_of_buffer l) (footprint h next (L.tl n)) /\\\n cells_pairwise_disjoint h next (L.tl n)\n\nval cells_live_freeable: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_live_freeable #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.live h l /\\ B.freeable l /\\\n cells_live_freeable h next (L.tl n)\n\nval invariant: #a:Type -> h:HS.mem -> l: t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\nlet invariant #a h l n =\n cells_live_freeable h l n /\\ cells_pairwise_disjoint h l n\n\n/// Normally this would be automatic and writing a custom lemma for that stuff\n/// would be needed only when writing an fsti, however, as stated earlier,\n/// recursion makes everything more difficult and this absolutely needs to be in\n/// scope otherwise the modifies-clause theory cannot deduce disjointness of\n/// fresh allocations w.r.t. the footprint.\nlet rec invariant_loc_in_footprint (#a: Type) (h: HS.mem) (l: t a) (n: list a): Lemma\n (requires well_formed h l n)\n (ensures B.loc_in (footprint h l n) h)\n (decreases n)\n [ SMTPat (invariant h l n) ]\n=\n if B.g_is_null l then\n ()\n else\n let next = (B.deref h l).next in\n invariant_loc_in_footprint h next (L.tl n)\n\n/// Another absolutely essential lemma, for which interestingly the pattern\n/// \"invariant h1 l n\" does not work, but well_formed does... I guess that's\n/// alright.\nlet rec frame (#a: Type) (l: t a) (n: list a) (r: B.loc) (h0 h1: HS.mem): Lemma\n (requires (\n well_formed h0 l n /\\\n invariant h0 l n /\\\n B.loc_disjoint r (footprint h0 l n) /\\\n B.modifies r h0 h1\n ))\n (ensures (\n well_formed h1 l n /\\\n footprint h1 l n == footprint h0 l n /\\\n cells h1 l n == cells h0 l n /\\\n invariant h1 l n\n ))\n (decreases n)\n [ SMTPat (well_formed h1 l n); SMTPat (B.modifies r h0 h1) ]\n=\n if B.g_is_null l then\n ()\n else\n frame (B.deref h0 l).next (L.tl n) r h0 h1\n\n/// Lemmas for working with linked lists\n/// ------------------------------------\n///\n/// Since it's a recursive data structure, a lot of the automated reasoning from\n/// LowStar.Buffer doesn't work since it requires inductions. Some helpful\n/// properties are therefore proved here.\n\nlet rec length_functional #a (h: HS.mem) (c: t a) (l1 l2: list a):\n Lemma\n (requires (well_formed h c l1 /\\ well_formed h c l2))\n (ensures (l1 == l2))\n (decreases (G.reveal l1))\n [ SMTPat (well_formed h c l1); SMTPat (well_formed h c l2) ] =\n if B.g_is_null c\n then ()\n else\n let { next=next } = B.get h c 0 in\n length_functional h next (G.hide (L.tl l1)) (G.hide (L.tl l2))\n\n/// The footprint is based on loc_addr_of_buffer so that we can write a free\n/// operation that operates on the footprint. However, this invalidates most of\n/// the helper lemmas for disjointness that were present in the previous\n/// iteration of this module, named LinkedList4.fst. It's ok, we bake\n/// disjointness directly onto the invariant and it works just as well.\n\n/// Connection between ``cells`` and ``v``\n/// --------------------------------------\n\n/// Redefining some helpers.\n#push-options \"--ifuel 1\"\nlet rec gmap #a #b (f: a -> GTot b) (xs: list a): GTot (list b) =\n match xs with\n | [] -> []\n | x :: xs -> f x :: gmap f xs\n\n/// \"As we all know\", right folds are the easiest to work with because they\n/// follow the natural structure of recursion (left-folds are evil).\nlet rec gfold_right #a #b (f: b -> a -> GTot b) (xs: list a) (acc: b): Ghost b\n (requires True)\n (ensures fun _ -> True)\n (decreases xs)\n=\n match xs with\n | [] -> acc\n | x :: xs -> f (gfold_right f xs acc) x\n#pop-options\n\n/// Connecting ``v`` and ``cells``.\n///\n/// See ``tests/Wireguard.fst`` for why going through gmap is important and how\n/// to structure your definitions (should you need to reason about ``cells``) in\n/// a way that will make your life easier.\nlet deref_data #a (h: HS.mem) (c: B.pointer (cell a)): GTot a =\n (B.deref h c).data\n\n#push-options \"--fuel 1 --ifuel 1\"\nlet rec deref_cells_is_v #a (h: HS.mem) (ll: t a) (l: list a): Lemma\n (requires\n well_formed h ll l /\\\n invariant h ll l)\n (ensures\n gmap (deref_data h) (cells h ll l) == l)\n (decreases l)\n [ SMTPat (well_formed h ll l) ]\n=\n if B.g_is_null ll then\n ()\n else\n deref_cells_is_v h (B.deref h ll).next (List.Tot.tl l)\n#pop-options\n\n/// Connecting ``footprint`` and ``cells``.\n///\n/// This allows conveniently switching to a low-level representation of the\n/// footprint in case clients want to do some ultra-precise reasoning about\n/// what's happening with the list spine. However, the fact that all\n/// operations from this module specify things in terms of ``cells`` along with\n/// ``same_cells_same_pointer`` should cover most common cases. (This predates\n/// the introduction of ``same_cells_same_pointer``.)\n///\n/// So, no SMTPat here. Note, I also used the same trick as above with\n/// higher-order combinators to isolate the use of the argument ``h`` to a\n/// single sub-term rather than having it passed down the recursive calls.\n\nPartial implementation:\nlet rec footprint_via_cells_is_footprint #a (h: HS.mem) (ll: t a) (l: list a)\n : Lemma (requires well_formed h ll l)\n (ensures\n footprint h ll l ==\n (gfold_right B.loc_union (gmap B.loc_addr_of_buffer (cells h ll l)) B.loc_none))\n (decreases l) =\n \n"
},
{
"role": "assistant",
"content": "if B.g_is_null ll\nthen ()\nelse footprint_via_cells_is_footprint h (B.deref h ll).next (List.Tot.tl l)"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: h: FStar.Monotonic.HyperStack.mem -> l: LowStar.Lib.LinkedList.t a -> n: Prims.list a\n -> Prims.Ghost Type0\nEffect: Prims.Ghost\n\nContext:\nmodule LowStar.Lib.LinkedList\n\n/// This module, oftentimes referred to as LL1 in other parts of the\n/// documentation, provides a low-level view over a linked list. Unless you\n/// intend on modifying the structure of the linked list (e.g. removing cells)\n/// or iterating over each cell, your are probably better off using LL2.\n///\n/// This module is intentionally not relying on a tight abstraction, to i) permit\n/// direct-style iteration by clients and ii) allow stating some SMTPat lemmas\n/// only on LL1 rather than replicating them in LL2 (going from LL2 to LL1 can\n/// happen automatically at the SMT level since LL2 does not have an fsti either).\n///\n/// Some open questions relative to the design of this module:\n/// - should ``v`` be defined directly in terms of ``cells`` rather than have a\n/// recursive definition along with an equivalence lemma?\n/// - should ``footprint`` be defined directly in terms of ``cells`` rather than\n/// have a recursive definition along with an equivalence lemma?\n/// - is it worth the trouble defining a footprint? my intuition is that it's\n/// fine and allows some clients to do very precise reasoning -- other modules\n/// (e.g. AssocList), when establishing abstraction boundaries, can simplify\n/// things and get rid of the precise footprint reasoning in favor of a static\n/// region-based footprint\n///\n/// Some remaining work items:\n/// - switch to C.Loops.while\n\nopen LowStar.BufferOps\n\nmodule B = LowStar.Buffer\nmodule HS = FStar.HyperStack\nmodule G = FStar.Ghost\nmodule L = FStar.List.Tot\nmodule U32 = FStar.UInt32\nmodule ST = FStar.HyperStack.ST\n\nopen FStar.HyperStack.ST\n\n/// Breaking from my own recommendation here... using fuel 1 / ifuel 1 since\n/// everything does recursive list reasoning.\n#set-options \"--__no_positivity --fuel 1 --ifuel 1\"\n\n/// Definition of the Low* type\n/// ---------------------------\nnoeq\ntype t (a: Type0) =\n B.pointer_or_null (cell a)\n\nand cell (a: Type0) = {\n next: t a;\n data: a;\n}\n\n/// Unlike the canonical style that associates a ``v`` function to any stateful\n/// representation, we use a relation here. Writing a function is just\n/// impossible, since we can't prove termination of ``v`` owing to the fact that\n/// there may be an infinite number of cells in the heap.\n///\n/// LL2 is the module that switches to the function style.\n///\n/// Note: no need to use erased here for ``l`` since the function is already in ``GTot``.\nlet rec well_formed #a (h: HS.mem) (c: t a) (l: list a): GTot Type0 (decreases l) =\n B.live h c /\\ (\n match l with\n | [] ->\n B.g_is_null c\n | a :: q ->\n B.length c == 1 /\\ (\n let { data=data; next=next } = B.get h c 0 in\n a == data /\\\n well_formed h next (G.hide q)\n ))\n\n/// Precise reasoning via the cells predicate\n/// -----------------------------------------\n///\n/// The modifies clauses are given either on the footprint of the list, or on a\n/// region that includes the footprint of the list (higher up the stack, e.g.\n/// LL2). This is good, but does not rule out \"stupid\" implementations that\n/// might re-allocate every single list cell when, say, doing a pop.\n///\n/// For advanced usages which take pointers directly to individual list cells,\n/// it's important to rule out these cases. We thus define the ``cells``\n/// predicate for that purpose.\n\nlet rec cells #a (h: HS.mem) (c: t a) (l: list a): Ghost (list (B.pointer (cell a)))\n (requires well_formed h c l)\n (ensures fun _ -> True)\n (decreases l)\n=\n if B.g_is_null c then\n []\n else\n c :: cells h (B.deref h c).next (List.Tot.tl l)\n\n/// A missing bit of automated reasoning\n/// ------------------------------------\n\n/// This should be absolutely trivial, yet, because there's no pattern on\n/// null_unique, we have to do the case analysis and the decomposition\n/// ourselves. This is quite tedious, as it usually takes a while (at least it\n/// did for me!) to realize that this property does not hold automatically.\n///\n/// I'm setting an SMTPat on this lemma because it's absolutely impossible to\n/// get any work done without it.\nlet same_cells_same_pointer #a (h0 h1: HS.mem) (ll0 ll1: t a) (l0 l1: list a): Lemma\n (requires (\n well_formed h0 ll0 l0 /\\\n well_formed h1 ll1 l1 /\\\n cells h0 ll0 l0 == cells h1 ll1 l1))\n (ensures (\n ll0 == ll1))\n [ SMTPat (cells h0 ll0 l0); SMTPat (cells h1 ll1 l1) ]\n=\n if B.g_is_null ll0 && B.g_is_null ll1 then begin\n B.null_unique ll0;\n B.null_unique ll1\n end else if not (B.g_is_null ll0) && not (B.g_is_null ll1) then begin\n ()\n end else\n false_elim ()\n\n/// Classic stateful reasoning lemmas\n/// ---------------------------------\n///\n/// This essentially follows ``stateful``, the type class of a mutable piece of\n/// state defined using the modifies-theory of Low*v2 and equipped with the\n/// right lemmas to enable usage by clients. See\n/// hacl-star/code/streaming/Hacl.Streaming.Interface.fst for more context.\n\nval footprint: (#a: Type) -> (h: HS.mem) -> (l: t a) -> (n: list a) -> Ghost B.loc\n (requires (well_formed h l n))\n (ensures (fun refs -> True))\n (decreases n)\n\nlet rec footprint #a h l n =\n if B.g_is_null l\n then B.loc_none\n else\n let {next = next} = B.get h l 0 in\n let refs = footprint h next (G.hide (L.tl n)) in\n B.loc_union (B.loc_addr_of_buffer l) refs\n\n/// Departing from LinkedList4 here. I prefer to bolt these into the invariant\n/// rather than requiring on some wizard lemmas from LowStar.Monotonic.Buffer to\n/// deduce these.\nval cells_pairwise_disjoint: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_pairwise_disjoint #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.loc_disjoint (B.loc_addr_of_buffer l) (footprint h next (L.tl n)) /\\\n cells_pairwise_disjoint h next (L.tl n)\n\nval cells_live_freeable: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nPartial implementation:\nlet rec cells_live_freeable #_ h l n =\n \n"
},
{
"role": "assistant",
"content": "if B.g_is_null l\nthen True\nelse\n let next = (B.deref h l).next in\n B.live h l /\\ B.freeable l /\\ cells_live_freeable h next (L.tl n)"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: gn: FStar.Ghost.erased (Prims.list a) -> l: LowStar.Lib.LinkedList.t a\n -> FStar.HyperStack.ST.Stack FStar.UInt32.t\nEffect: FStar.HyperStack.ST.Stack\n\nContext:\nmodule LowStar.Lib.LinkedList\n\n/// This module, oftentimes referred to as LL1 in other parts of the\n/// documentation, provides a low-level view over a linked list. Unless you\n/// intend on modifying the structure of the linked list (e.g. removing cells)\n/// or iterating over each cell, your are probably better off using LL2.\n///\n/// This module is intentionally not relying on a tight abstraction, to i) permit\n/// direct-style iteration by clients and ii) allow stating some SMTPat lemmas\n/// only on LL1 rather than replicating them in LL2 (going from LL2 to LL1 can\n/// happen automatically at the SMT level since LL2 does not have an fsti either).\n///\n/// Some open questions relative to the design of this module:\n/// - should ``v`` be defined directly in terms of ``cells`` rather than have a\n/// recursive definition along with an equivalence lemma?\n/// - should ``footprint`` be defined directly in terms of ``cells`` rather than\n/// have a recursive definition along with an equivalence lemma?\n/// - is it worth the trouble defining a footprint? my intuition is that it's\n/// fine and allows some clients to do very precise reasoning -- other modules\n/// (e.g. AssocList), when establishing abstraction boundaries, can simplify\n/// things and get rid of the precise footprint reasoning in favor of a static\n/// region-based footprint\n///\n/// Some remaining work items:\n/// - switch to C.Loops.while\n\nopen LowStar.BufferOps\n\nmodule B = LowStar.Buffer\nmodule HS = FStar.HyperStack\nmodule G = FStar.Ghost\nmodule L = FStar.List.Tot\nmodule U32 = FStar.UInt32\nmodule ST = FStar.HyperStack.ST\n\nopen FStar.HyperStack.ST\n\n/// Breaking from my own recommendation here... using fuel 1 / ifuel 1 since\n/// everything does recursive list reasoning.\n#set-options \"--__no_positivity --fuel 1 --ifuel 1\"\n\n/// Definition of the Low* type\n/// ---------------------------\nnoeq\ntype t (a: Type0) =\n B.pointer_or_null (cell a)\n\nand cell (a: Type0) = {\n next: t a;\n data: a;\n}\n\n/// Unlike the canonical style that associates a ``v`` function to any stateful\n/// representation, we use a relation here. Writing a function is just\n/// impossible, since we can't prove termination of ``v`` owing to the fact that\n/// there may be an infinite number of cells in the heap.\n///\n/// LL2 is the module that switches to the function style.\n///\n/// Note: no need to use erased here for ``l`` since the function is already in ``GTot``.\nlet rec well_formed #a (h: HS.mem) (c: t a) (l: list a): GTot Type0 (decreases l) =\n B.live h c /\\ (\n match l with\n | [] ->\n B.g_is_null c\n | a :: q ->\n B.length c == 1 /\\ (\n let { data=data; next=next } = B.get h c 0 in\n a == data /\\\n well_formed h next (G.hide q)\n ))\n\n/// Precise reasoning via the cells predicate\n/// -----------------------------------------\n///\n/// The modifies clauses are given either on the footprint of the list, or on a\n/// region that includes the footprint of the list (higher up the stack, e.g.\n/// LL2). This is good, but does not rule out \"stupid\" implementations that\n/// might re-allocate every single list cell when, say, doing a pop.\n///\n/// For advanced usages which take pointers directly to individual list cells,\n/// it's important to rule out these cases. We thus define the ``cells``\n/// predicate for that purpose.\n\nlet rec cells #a (h: HS.mem) (c: t a) (l: list a): Ghost (list (B.pointer (cell a)))\n (requires well_formed h c l)\n (ensures fun _ -> True)\n (decreases l)\n=\n if B.g_is_null c then\n []\n else\n c :: cells h (B.deref h c).next (List.Tot.tl l)\n\n/// A missing bit of automated reasoning\n/// ------------------------------------\n\n/// This should be absolutely trivial, yet, because there's no pattern on\n/// null_unique, we have to do the case analysis and the decomposition\n/// ourselves. This is quite tedious, as it usually takes a while (at least it\n/// did for me!) to realize that this property does not hold automatically.\n///\n/// I'm setting an SMTPat on this lemma because it's absolutely impossible to\n/// get any work done without it.\nlet same_cells_same_pointer #a (h0 h1: HS.mem) (ll0 ll1: t a) (l0 l1: list a): Lemma\n (requires (\n well_formed h0 ll0 l0 /\\\n well_formed h1 ll1 l1 /\\\n cells h0 ll0 l0 == cells h1 ll1 l1))\n (ensures (\n ll0 == ll1))\n [ SMTPat (cells h0 ll0 l0); SMTPat (cells h1 ll1 l1) ]\n=\n if B.g_is_null ll0 && B.g_is_null ll1 then begin\n B.null_unique ll0;\n B.null_unique ll1\n end else if not (B.g_is_null ll0) && not (B.g_is_null ll1) then begin\n ()\n end else\n false_elim ()\n\n/// Classic stateful reasoning lemmas\n/// ---------------------------------\n///\n/// This essentially follows ``stateful``, the type class of a mutable piece of\n/// state defined using the modifies-theory of Low*v2 and equipped with the\n/// right lemmas to enable usage by clients. See\n/// hacl-star/code/streaming/Hacl.Streaming.Interface.fst for more context.\n\nval footprint: (#a: Type) -> (h: HS.mem) -> (l: t a) -> (n: list a) -> Ghost B.loc\n (requires (well_formed h l n))\n (ensures (fun refs -> True))\n (decreases n)\n\nlet rec footprint #a h l n =\n if B.g_is_null l\n then B.loc_none\n else\n let {next = next} = B.get h l 0 in\n let refs = footprint h next (G.hide (L.tl n)) in\n B.loc_union (B.loc_addr_of_buffer l) refs\n\n/// Departing from LinkedList4 here. I prefer to bolt these into the invariant\n/// rather than requiring on some wizard lemmas from LowStar.Monotonic.Buffer to\n/// deduce these.\nval cells_pairwise_disjoint: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_pairwise_disjoint #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.loc_disjoint (B.loc_addr_of_buffer l) (footprint h next (L.tl n)) /\\\n cells_pairwise_disjoint h next (L.tl n)\n\nval cells_live_freeable: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_live_freeable #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.live h l /\\ B.freeable l /\\\n cells_live_freeable h next (L.tl n)\n\nval invariant: #a:Type -> h:HS.mem -> l: t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\nlet invariant #a h l n =\n cells_live_freeable h l n /\\ cells_pairwise_disjoint h l n\n\n/// Normally this would be automatic and writing a custom lemma for that stuff\n/// would be needed only when writing an fsti, however, as stated earlier,\n/// recursion makes everything more difficult and this absolutely needs to be in\n/// scope otherwise the modifies-clause theory cannot deduce disjointness of\n/// fresh allocations w.r.t. the footprint.\nlet rec invariant_loc_in_footprint (#a: Type) (h: HS.mem) (l: t a) (n: list a): Lemma\n (requires well_formed h l n)\n (ensures B.loc_in (footprint h l n) h)\n (decreases n)\n [ SMTPat (invariant h l n) ]\n=\n if B.g_is_null l then\n ()\n else\n let next = (B.deref h l).next in\n invariant_loc_in_footprint h next (L.tl n)\n\n/// Another absolutely essential lemma, for which interestingly the pattern\n/// \"invariant h1 l n\" does not work, but well_formed does... I guess that's\n/// alright.\nlet rec frame (#a: Type) (l: t a) (n: list a) (r: B.loc) (h0 h1: HS.mem): Lemma\n (requires (\n well_formed h0 l n /\\\n invariant h0 l n /\\\n B.loc_disjoint r (footprint h0 l n) /\\\n B.modifies r h0 h1\n ))\n (ensures (\n well_formed h1 l n /\\\n footprint h1 l n == footprint h0 l n /\\\n cells h1 l n == cells h0 l n /\\\n invariant h1 l n\n ))\n (decreases n)\n [ SMTPat (well_formed h1 l n); SMTPat (B.modifies r h0 h1) ]\n=\n if B.g_is_null l then\n ()\n else\n frame (B.deref h0 l).next (L.tl n) r h0 h1\n\n/// Lemmas for working with linked lists\n/// ------------------------------------\n///\n/// Since it's a recursive data structure, a lot of the automated reasoning from\n/// LowStar.Buffer doesn't work since it requires inductions. Some helpful\n/// properties are therefore proved here.\n\nlet rec length_functional #a (h: HS.mem) (c: t a) (l1 l2: list a):\n Lemma\n (requires (well_formed h c l1 /\\ well_formed h c l2))\n (ensures (l1 == l2))\n (decreases (G.reveal l1))\n [ SMTPat (well_formed h c l1); SMTPat (well_formed h c l2) ] =\n if B.g_is_null c\n then ()\n else\n let { next=next } = B.get h c 0 in\n length_functional h next (G.hide (L.tl l1)) (G.hide (L.tl l2))\n\n/// The footprint is based on loc_addr_of_buffer so that we can write a free\n/// operation that operates on the footprint. However, this invalidates most of\n/// the helper lemmas for disjointness that were present in the previous\n/// iteration of this module, named LinkedList4.fst. It's ok, we bake\n/// disjointness directly onto the invariant and it works just as well.\n\n/// Connection between ``cells`` and ``v``\n/// --------------------------------------\n\n/// Redefining some helpers.\n#push-options \"--ifuel 1\"\nlet rec gmap #a #b (f: a -> GTot b) (xs: list a): GTot (list b) =\n match xs with\n | [] -> []\n | x :: xs -> f x :: gmap f xs\n\n/// \"As we all know\", right folds are the easiest to work with because they\n/// follow the natural structure of recursion (left-folds are evil).\nlet rec gfold_right #a #b (f: b -> a -> GTot b) (xs: list a) (acc: b): Ghost b\n (requires True)\n (ensures fun _ -> True)\n (decreases xs)\n=\n match xs with\n | [] -> acc\n | x :: xs -> f (gfold_right f xs acc) x\n#pop-options\n\n/// Connecting ``v`` and ``cells``.\n///\n/// See ``tests/Wireguard.fst`` for why going through gmap is important and how\n/// to structure your definitions (should you need to reason about ``cells``) in\n/// a way that will make your life easier.\nlet deref_data #a (h: HS.mem) (c: B.pointer (cell a)): GTot a =\n (B.deref h c).data\n\n#push-options \"--fuel 1 --ifuel 1\"\nlet rec deref_cells_is_v #a (h: HS.mem) (ll: t a) (l: list a): Lemma\n (requires\n well_formed h ll l /\\\n invariant h ll l)\n (ensures\n gmap (deref_data h) (cells h ll l) == l)\n (decreases l)\n [ SMTPat (well_formed h ll l) ]\n=\n if B.g_is_null ll then\n ()\n else\n deref_cells_is_v h (B.deref h ll).next (List.Tot.tl l)\n#pop-options\n\n/// Connecting ``footprint`` and ``cells``.\n///\n/// This allows conveniently switching to a low-level representation of the\n/// footprint in case clients want to do some ultra-precise reasoning about\n/// what's happening with the list spine. However, the fact that all\n/// operations from this module specify things in terms of ``cells`` along with\n/// ``same_cells_same_pointer`` should cover most common cases. (This predates\n/// the introduction of ``same_cells_same_pointer``.)\n///\n/// So, no SMTPat here. Note, I also used the same trick as above with\n/// higher-order combinators to isolate the use of the argument ``h`` to a\n/// single sub-term rather than having it passed down the recursive calls.\n#push-options \"--fuel 2 --ifuel 2\"\nlet rec footprint_via_cells_is_footprint #a (h: HS.mem) (ll: t a) (l: list a): Lemma\n (requires well_formed h ll l)\n (ensures\n footprint h ll l == (gfold_right B.loc_union (gmap B.loc_addr_of_buffer (cells h ll l)) B.loc_none))\n (decreases l)\n=\n if B.g_is_null ll then\n ()\n else\n footprint_via_cells_is_footprint h (B.deref h ll).next (List.Tot.tl l)\n#pop-options\n\n\n/// Stateful operations\n/// -------------------\n///\n/// One thing to note: we pass the region `r` explicitly, which is certainly not\n/// very convenient, but that's alright, another layer of abstraction will take\n/// care of adding existentials and extra machinery to make this more pleasant\n/// to use.\n///\n/// All of the operations below are low-level (in the sense that they rely on\n/// the predicate). I expect clients to use exclusively the variants of these\n/// functions present in LL2.\n\nval push: (#a: Type) -> (r: HS.rid) -> (n: G.erased (list a)) -> (pl: B.pointer (t a)) -> (x: a) ->\n ST unit\n (requires (fun h ->\n let l = B.deref h pl in\n B.live h pl /\\\n well_formed h l n /\\\n invariant h l n /\\\n ST.is_eternal_region r /\\\n B.(loc_includes (loc_region_only true r) (footprint h l n)) /\\\n B.(loc_disjoint (loc_buffer pl) (loc_region_only true r))\n ))\n (ensures (fun h0 _ h1 ->\n let n' = G.hide (x :: G.reveal n) in\n let l = B.deref h1 pl in\n // Style note: I don't repeat ``B.live pl`` in the post-condition since\n // ``B.modifies (loc_buffer pl) h0 h1`` implies that ``B.live h1 pl``.\n B.modifies (B.loc_buffer pl) h0 h1 /\\\n well_formed h1 l n' /\\\n invariant h1 l n' /\\\n B.(loc_includes (loc_region_only true r) (footprint h1 l n') /\\\n Cons? (cells h1 l n') /\\ List.Tot.tail (cells h1 l n') == cells h0 (B.deref h0 pl) n /\\\n B.fresh_loc (B.loc_addr_of_buffer (List.Tot.hd (cells h1 l n'))) h0 h1)\n ))\n\nlet push #a r n pl x =\n (**) let h0 = ST.get () in\n let l = !* pl in\n let c = { data = x; next = l } in\n\n let pc: B.pointer (cell a) = B.malloc r c 1ul in\n (**) let h1 = ST.get () in\n (**) B.(modifies_only_not_unused_in loc_none h0 h1);\n (**) assert B.(loc_disjoint (loc_buffer pc) (footprint h0 l n));\n\n pl *= pc;\n (**) let h2 = ST.get () in\n (**) let n' = G.hide (x :: G.reveal n) in\n (**) B.(modifies_trans loc_none h0 h1 (loc_buffer pl) h2);\n (**) assert (well_formed h2 (B.deref h2 pl) n');\n (**) assert (invariant h2 (B.deref h2 pl) n');\n (**) assert ((B.deref h2 (B.deref h2 pl)).next == l);\n\n ()\n\nval pop: (#a: Type) -> (r: HS.rid) -> (n: G.erased (list a)) -> (pl: B.pointer (t a)) ->\n ST a\n (requires (fun h ->\n let l = B.deref h pl in\n B.live h pl /\\\n well_formed h l n /\\\n invariant h l n /\\\n ST.is_eternal_region r /\\\n B.(loc_includes (loc_region_only true r) (footprint h l n)) /\\\n B.loc_disjoint (B.loc_buffer pl) (footprint h l n) /\\\n Cons? n\n ))\n (ensures (fun h0 x h1 ->\n let l = B.deref h1 pl in\n let n' = L.tl n in\n x == L.hd n /\\\n // Introducing a super precise modifies clause (e.g. loc_addr_of_buffer\n // (B.deref h0 pl)) here is not useful and prevents trigger-based\n // reasoning, while also requiring clients to unroll footprint. There's a\n // tension here between revealing that ``pop`` does nothing stupid (e.g.\n // re-allocating all the list cells) and providing an abstract enough\n // predicate to work with.\n B.(modifies (loc_buffer pl `loc_union` footprint h0 (B.deref h0 pl) n) h0 h1) /\\\n well_formed h1 l n' /\\\n invariant h1 l n' /\\\n B.(loc_includes (loc_region_only true r) (footprint h1 l n')) /\\\n List.Tot.tail (cells h0 (B.deref h0 pl) n) == cells h1 l n'))\n\nlet pop #a r n pl =\n let l = !* pl in\n let r = (!*l).data in\n let next = (!*l).next in\n (**) let h1 = ST.get () in\n (**) assert (cells h1 l n == l :: cells h1 next (List.Tot.tl n));\n (**) assert B.(loc_disjoint (loc_buffer pl) (footprint h1 l n));\n (**) assert B.(loc_disjoint (loc_buffer pl) (footprint h1 next (List.Tot.tl n)));\n\n pl *= next;\n B.free l;\n r\n\nval free_: (#a: Type) -> (#n: G.erased (list a)) -> (l: t a) ->\n ST unit\n (requires (fun h ->\n well_formed h l n /\\\n invariant h l n\n ))\n (ensures (fun h0 _ h1 ->\n B.(modifies (footprint h0 l n) h0 h1)))\n\nlet rec free_ #a #n l =\n if B.is_null l then\n ()\n else begin\n let tl: G.erased (list a) = G.hide (L.tl (G.reveal n)) in\n free_ #_ #tl (!*l).next;\n B.free l\n end\n\nval free: (#a: Type) -> (#n: G.erased (list a)) -> (pl: B.pointer (t a)) ->\n ST unit\n (requires (fun h ->\n let l = B.deref h pl in\n B.live h pl /\\\n well_formed h l n /\\\n invariant h l n /\\\n B.loc_disjoint (B.loc_buffer pl) (footprint h l n)\n ))\n (ensures (fun h0 _ h1 ->\n let l = B.deref h1 pl in\n well_formed h1 l [] /\\\n invariant h1 l [] /\\\n footprint h1 l [] == B.loc_none /\\\n cells h1 l [] == [] /\\\n B.(modifies (footprint h0 (B.deref h0 pl) n `loc_union` loc_buffer pl) h0 h1)))\n\nlet free #a #n pl =\n free_ #_ #n !*pl;\n pl *= B.null\n\nval length (#a: Type) (gn: G.erased (list a)) (l: t a): Stack UInt32.t\n (requires (fun h -> well_formed h l gn))\n (ensures (fun h0 n h1 ->\n h0 == h1 /\\\n U32.v n = L.length (G.reveal gn)\n ))\n\nPartial implementation:\nlet rec length #a gn l =\n \n"
},
{
"role": "assistant",
"content": "if B.is_null l\nthen 0ul\nelse\n let open U32 in\n let c = !*l in\n let next = c.next in\n let n = length (G.hide (L.tail (G.reveal gn))) next in\n if n = 0xfffffffful\n then LowStar.Failure.failwith \"Integer overflow in LowStar.LinkedList.length\"\n else n +^ 1ul"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: f: (_: a -> Prims.GTot b) -> xs: Prims.list a -> Prims.GTot (Prims.list b)\nEffect: Prims.GTot\n\nContext:\nmodule LowStar.Lib.LinkedList\n\n/// This module, oftentimes referred to as LL1 in other parts of the\n/// documentation, provides a low-level view over a linked list. Unless you\n/// intend on modifying the structure of the linked list (e.g. removing cells)\n/// or iterating over each cell, your are probably better off using LL2.\n///\n/// This module is intentionally not relying on a tight abstraction, to i) permit\n/// direct-style iteration by clients and ii) allow stating some SMTPat lemmas\n/// only on LL1 rather than replicating them in LL2 (going from LL2 to LL1 can\n/// happen automatically at the SMT level since LL2 does not have an fsti either).\n///\n/// Some open questions relative to the design of this module:\n/// - should ``v`` be defined directly in terms of ``cells`` rather than have a\n/// recursive definition along with an equivalence lemma?\n/// - should ``footprint`` be defined directly in terms of ``cells`` rather than\n/// have a recursive definition along with an equivalence lemma?\n/// - is it worth the trouble defining a footprint? my intuition is that it's\n/// fine and allows some clients to do very precise reasoning -- other modules\n/// (e.g. AssocList), when establishing abstraction boundaries, can simplify\n/// things and get rid of the precise footprint reasoning in favor of a static\n/// region-based footprint\n///\n/// Some remaining work items:\n/// - switch to C.Loops.while\n\nopen LowStar.BufferOps\n\nmodule B = LowStar.Buffer\nmodule HS = FStar.HyperStack\nmodule G = FStar.Ghost\nmodule L = FStar.List.Tot\nmodule U32 = FStar.UInt32\nmodule ST = FStar.HyperStack.ST\n\nopen FStar.HyperStack.ST\n\n/// Breaking from my own recommendation here... using fuel 1 / ifuel 1 since\n/// everything does recursive list reasoning.\n#set-options \"--__no_positivity --fuel 1 --ifuel 1\"\n\n/// Definition of the Low* type\n/// ---------------------------\nnoeq\ntype t (a: Type0) =\n B.pointer_or_null (cell a)\n\nand cell (a: Type0) = {\n next: t a;\n data: a;\n}\n\n/// Unlike the canonical style that associates a ``v`` function to any stateful\n/// representation, we use a relation here. Writing a function is just\n/// impossible, since we can't prove termination of ``v`` owing to the fact that\n/// there may be an infinite number of cells in the heap.\n///\n/// LL2 is the module that switches to the function style.\n///\n/// Note: no need to use erased here for ``l`` since the function is already in ``GTot``.\nlet rec well_formed #a (h: HS.mem) (c: t a) (l: list a): GTot Type0 (decreases l) =\n B.live h c /\\ (\n match l with\n | [] ->\n B.g_is_null c\n | a :: q ->\n B.length c == 1 /\\ (\n let { data=data; next=next } = B.get h c 0 in\n a == data /\\\n well_formed h next (G.hide q)\n ))\n\n/// Precise reasoning via the cells predicate\n/// -----------------------------------------\n///\n/// The modifies clauses are given either on the footprint of the list, or on a\n/// region that includes the footprint of the list (higher up the stack, e.g.\n/// LL2). This is good, but does not rule out \"stupid\" implementations that\n/// might re-allocate every single list cell when, say, doing a pop.\n///\n/// For advanced usages which take pointers directly to individual list cells,\n/// it's important to rule out these cases. We thus define the ``cells``\n/// predicate for that purpose.\n\nlet rec cells #a (h: HS.mem) (c: t a) (l: list a): Ghost (list (B.pointer (cell a)))\n (requires well_formed h c l)\n (ensures fun _ -> True)\n (decreases l)\n=\n if B.g_is_null c then\n []\n else\n c :: cells h (B.deref h c).next (List.Tot.tl l)\n\n/// A missing bit of automated reasoning\n/// ------------------------------------\n\n/// This should be absolutely trivial, yet, because there's no pattern on\n/// null_unique, we have to do the case analysis and the decomposition\n/// ourselves. This is quite tedious, as it usually takes a while (at least it\n/// did for me!) to realize that this property does not hold automatically.\n///\n/// I'm setting an SMTPat on this lemma because it's absolutely impossible to\n/// get any work done without it.\nlet same_cells_same_pointer #a (h0 h1: HS.mem) (ll0 ll1: t a) (l0 l1: list a): Lemma\n (requires (\n well_formed h0 ll0 l0 /\\\n well_formed h1 ll1 l1 /\\\n cells h0 ll0 l0 == cells h1 ll1 l1))\n (ensures (\n ll0 == ll1))\n [ SMTPat (cells h0 ll0 l0); SMTPat (cells h1 ll1 l1) ]\n=\n if B.g_is_null ll0 && B.g_is_null ll1 then begin\n B.null_unique ll0;\n B.null_unique ll1\n end else if not (B.g_is_null ll0) && not (B.g_is_null ll1) then begin\n ()\n end else\n false_elim ()\n\n/// Classic stateful reasoning lemmas\n/// ---------------------------------\n///\n/// This essentially follows ``stateful``, the type class of a mutable piece of\n/// state defined using the modifies-theory of Low*v2 and equipped with the\n/// right lemmas to enable usage by clients. See\n/// hacl-star/code/streaming/Hacl.Streaming.Interface.fst for more context.\n\nval footprint: (#a: Type) -> (h: HS.mem) -> (l: t a) -> (n: list a) -> Ghost B.loc\n (requires (well_formed h l n))\n (ensures (fun refs -> True))\n (decreases n)\n\nlet rec footprint #a h l n =\n if B.g_is_null l\n then B.loc_none\n else\n let {next = next} = B.get h l 0 in\n let refs = footprint h next (G.hide (L.tl n)) in\n B.loc_union (B.loc_addr_of_buffer l) refs\n\n/// Departing from LinkedList4 here. I prefer to bolt these into the invariant\n/// rather than requiring on some wizard lemmas from LowStar.Monotonic.Buffer to\n/// deduce these.\nval cells_pairwise_disjoint: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_pairwise_disjoint #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.loc_disjoint (B.loc_addr_of_buffer l) (footprint h next (L.tl n)) /\\\n cells_pairwise_disjoint h next (L.tl n)\n\nval cells_live_freeable: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_live_freeable #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.live h l /\\ B.freeable l /\\\n cells_live_freeable h next (L.tl n)\n\nval invariant: #a:Type -> h:HS.mem -> l: t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\nlet invariant #a h l n =\n cells_live_freeable h l n /\\ cells_pairwise_disjoint h l n\n\n/// Normally this would be automatic and writing a custom lemma for that stuff\n/// would be needed only when writing an fsti, however, as stated earlier,\n/// recursion makes everything more difficult and this absolutely needs to be in\n/// scope otherwise the modifies-clause theory cannot deduce disjointness of\n/// fresh allocations w.r.t. the footprint.\nlet rec invariant_loc_in_footprint (#a: Type) (h: HS.mem) (l: t a) (n: list a): Lemma\n (requires well_formed h l n)\n (ensures B.loc_in (footprint h l n) h)\n (decreases n)\n [ SMTPat (invariant h l n) ]\n=\n if B.g_is_null l then\n ()\n else\n let next = (B.deref h l).next in\n invariant_loc_in_footprint h next (L.tl n)\n\n/// Another absolutely essential lemma, for which interestingly the pattern\n/// \"invariant h1 l n\" does not work, but well_formed does... I guess that's\n/// alright.\nlet rec frame (#a: Type) (l: t a) (n: list a) (r: B.loc) (h0 h1: HS.mem): Lemma\n (requires (\n well_formed h0 l n /\\\n invariant h0 l n /\\\n B.loc_disjoint r (footprint h0 l n) /\\\n B.modifies r h0 h1\n ))\n (ensures (\n well_formed h1 l n /\\\n footprint h1 l n == footprint h0 l n /\\\n cells h1 l n == cells h0 l n /\\\n invariant h1 l n\n ))\n (decreases n)\n [ SMTPat (well_formed h1 l n); SMTPat (B.modifies r h0 h1) ]\n=\n if B.g_is_null l then\n ()\n else\n frame (B.deref h0 l).next (L.tl n) r h0 h1\n\n/// Lemmas for working with linked lists\n/// ------------------------------------\n///\n/// Since it's a recursive data structure, a lot of the automated reasoning from\n/// LowStar.Buffer doesn't work since it requires inductions. Some helpful\n/// properties are therefore proved here.\n\nlet rec length_functional #a (h: HS.mem) (c: t a) (l1 l2: list a):\n Lemma\n (requires (well_formed h c l1 /\\ well_formed h c l2))\n (ensures (l1 == l2))\n (decreases (G.reveal l1))\n [ SMTPat (well_formed h c l1); SMTPat (well_formed h c l2) ] =\n if B.g_is_null c\n then ()\n else\n let { next=next } = B.get h c 0 in\n length_functional h next (G.hide (L.tl l1)) (G.hide (L.tl l2))\n\n/// The footprint is based on loc_addr_of_buffer so that we can write a free\n/// operation that operates on the footprint. However, this invalidates most of\n/// the helper lemmas for disjointness that were present in the previous\n/// iteration of this module, named LinkedList4.fst. It's ok, we bake\n/// disjointness directly onto the invariant and it works just as well.\n\n/// Connection between ``cells`` and ``v``\n/// --------------------------------------\n\n/// Redefining some helpers.\n\nPartial implementation:\nlet rec gmap #a #b (f: (a -> GTot b)) (xs: list a) : GTot (list b) =\n \n"
},
{
"role": "assistant",
"content": "match xs with\n| [] -> []\n| x :: xs -> f x :: gmap f xs"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n ke: Hacl.Bignum.Exponentiation.exp t ->\n a: Spec.Hash.Definitions.hash_alg{Spec.RSAPSS.hash_is_supported a} ->\n modBits: Hacl.Impl.RSAPSS.modBits_t t ->\n rsapss_load_pkey: Hacl.Impl.RSAPSS.Keys.rsapss_load_pkey_st t ke modBits ->\n rsapss_verify: Hacl.Impl.RSAPSS.rsapss_verify_st t ke a modBits\n -> Hacl.Impl.RSAPSS.rsapss_pkey_verify_st t ke a modBits\nEffect: Prims.Tot\n\nContext:\nmodule Hacl.Impl.RSAPSS\n\nopen FStar.HyperStack\nopen FStar.HyperStack.ST\nopen FStar.Mul\n\nopen Lib.IntTypes\nopen Lib.Buffer\n\nopen Hacl.Bignum.Definitions\n\nmodule ST = FStar.HyperStack.ST\n\nmodule Hash = Spec.Agile.Hash\nmodule SB = Hacl.Spec.Bignum\nmodule BB = Hacl.Spec.Bignum.Base\nmodule SD = Hacl.Spec.Bignum.Definitions\nmodule SM = Hacl.Spec.Bignum.Montgomery\nmodule SE = Hacl.Spec.Bignum.Exponentiation\n\nmodule BN = Hacl.Bignum\nmodule BE = Hacl.Bignum.Exponentiation\nmodule BM = Hacl.Bignum.Montgomery\n\nmodule S = Spec.RSAPSS\nmodule LS = Hacl.Spec.RSAPSS\nmodule LSeq = Lib.Sequence\n\nmodule RP = Hacl.Impl.RSAPSS.Padding\nmodule RM = Hacl.Impl.RSAPSS.MGF\nmodule RK = Hacl.Impl.RSAPSS.Keys\n\n#reset-options \"--z3rlimit 150 --fuel 0 --ifuel 0\"\n\ninline_for_extraction noextract\nlet modBits_t (t:limb_t) = modBits:size_t{1 < v modBits /\\ 2 * bits t * SD.blocks (v modBits) (bits t) <= max_size_t}\n\n\ninline_for_extraction noextract\nlet rsapss_sign_bn_st (t:limb_t) (ke:BE.exp t) (modBits:modBits_t t) =\n let len = blocks modBits (size (bits t)) in\n eBits:size_t\n -> dBits:size_t{LS.skey_len_pre t (v modBits) (v eBits) (v dBits)}\n -> skey:lbignum t (2ul *! len +! blocks eBits (size (bits t)) +! blocks dBits (size (bits t)))\n -> m:lbignum t len\n -> m':lbignum t len\n -> s:lbignum t len ->\n Stack bool\n (requires fun h -> len == ke.BE.bn.BN.len /\\\n live h skey /\\ live h m /\\ live h s /\\ live h m' /\\\n disjoint s m /\\ disjoint s skey /\\ disjoint m skey /\\\n disjoint m m' /\\ disjoint m' s /\\ disjoint m' skey /\\\n LS.rsapss_skey_pre (v modBits) (v eBits) (v dBits) (as_seq h skey) /\\\n bn_v h m < bn_v h (gsub skey 0ul len))\n (ensures fun h0 r h1 -> modifies (loc s |+| loc m') h0 h1 /\\\n (r, as_seq h1 s) == LS.rsapss_sign_bn (v modBits) (v eBits) (v dBits) (as_seq h0 skey) (as_seq h0 m))\n\n\ninline_for_extraction noextract\nval rsapss_sign_bn: #t:limb_t -> ke:BE.exp t -> modBits:modBits_t t -> rsapss_sign_bn_st t ke modBits\nlet rsapss_sign_bn #t ke modBits eBits dBits skey m m' s =\n [@inline_let] let bits : size_pos = bits t in\n let nLen = blocks modBits (size bits) in\n let eLen = blocks eBits (size bits) in\n let dLen = blocks dBits (size bits) in\n\n let n = sub skey 0ul nLen in\n let r2 = sub skey nLen nLen in\n let e = sub skey (nLen +! nLen) eLen in\n let d = sub skey (nLen +! nLen +! eLen) dLen in\n\n Math.Lemmas.pow2_le_compat (bits * v nLen) (v modBits);\n let h0 = ST.get () in\n SM.bn_precomp_r2_mod_n_lemma (v modBits - 1) (as_seq h0 n);\n BE.mk_bn_mod_exp_precompr2 nLen ke.BE.exp_ct_precomp n r2 m dBits d s;\n BE.mk_bn_mod_exp_precompr2 nLen ke.BE.exp_vt_precomp n r2 s eBits e m';\n let h1 = ST.get () in\n SD.bn_eval_inj (v nLen) (as_seq h1 s)\n (SE.bn_mod_exp_consttime_precompr2 (v nLen) (as_seq h0 n) (as_seq h0 r2)\n (as_seq h0 m) (v dBits) (as_seq h0 d));\n SD.bn_eval_inj (v nLen) (as_seq h1 m')\n (SE.bn_mod_exp_vartime_precompr2 (v nLen) (as_seq h0 n) (as_seq h0 r2)\n (as_seq h1 s) (v eBits) (as_seq h0 e));\n let eq_m = BN.bn_eq_mask nLen m m' in\n mapT nLen s (logand eq_m) s;\n BB.unsafe_bool_of_limb eq_m\n\n\ninline_for_extraction noextract\nlet rsapss_sign_msg_to_bn_st (t:limb_t) (a:Hash.hash_alg{S.hash_is_supported a}) (modBits:modBits_t t) =\n let len = blocks modBits (size (bits t)) in\n saltLen:size_t\n -> salt:lbuffer uint8 saltLen\n -> msgLen:size_t\n -> msg:lbuffer uint8 msgLen\n -> m:lbignum t len ->\n Stack unit\n (requires fun h ->\n live h salt /\\ live h msg /\\ live h m /\\\n disjoint salt msg /\\ disjoint m msg /\\ disjoint m salt /\\\n as_seq h m == LSeq.create (v len) (uint #t 0) /\\\n LS.rsapss_sign_pre a (v modBits) (v saltLen) (as_seq h salt) (v msgLen) (as_seq h msg))\n (ensures fun h0 _ h1 -> modifies (loc m) h0 h1 /\\\n as_seq h1 m == LS.rsapss_sign_msg_to_bn a (v modBits) (v saltLen) (as_seq h0 salt) (v msgLen) (as_seq h0 msg))\n\n\ninline_for_extraction noextract\nval rsapss_sign_msg_to_bn:\n #t:limb_t\n -> a:Hash.hash_alg{S.hash_is_supported a}\n -> modBits:modBits_t t ->\n rsapss_sign_msg_to_bn_st t a modBits\n\nlet rsapss_sign_msg_to_bn #t a modBits saltLen salt msgLen msg m =\n push_frame ();\n [@inline_let] let bits : size_pos = bits t in\n [@inline_let] let numb : size_pos = numbytes t in\n let nLen = blocks modBits (size bits) in\n\n let emBits = modBits -! 1ul in\n let emLen = blocks emBits 8ul in\n [@inline_let] let mLen = blocks emLen (size numb) in\n\n let em = create emLen (u8 0) in\n RP.pss_encode a saltLen salt msgLen msg emBits em;\n LS.blocks_bits_lemma t (v emBits);\n LS.blocks_numb_lemma t (v emBits);\n assert (SD.blocks (v emBits) bits = v mLen);\n assert (numb * v mLen <= max_size_t);\n assert (v mLen <= v nLen);\n\n let h' = ST.get () in\n update_sub_f h' m 0ul mLen\n (fun h -> SB.bn_from_bytes_be (v emLen) (as_seq h' em))\n (fun _ -> BN.bn_from_bytes_be emLen em (sub m 0ul mLen));\n pop_frame ()\n\n\ninline_for_extraction noextract\nlet rsapss_sign_compute_sgnt_st (t:limb_t) (ke:BE.exp t) (modBits:modBits_t t) =\n let len = blocks modBits (size (bits t)) in\n eBits:size_t\n -> dBits:size_t{LS.skey_len_pre t (v modBits) (v eBits) (v dBits)}\n -> skey:lbignum t (2ul *! len +! blocks eBits (size (bits t)) +! blocks dBits (size (bits t)))\n -> m:lbignum t len\n -> sgnt:lbuffer uint8 (blocks modBits 8ul) ->\n Stack bool\n (requires fun h -> len == ke.BE.bn.BN.len /\\\n live h sgnt /\\ live h skey /\\ live h m /\\\n disjoint sgnt skey /\\ disjoint m sgnt /\\ disjoint m skey /\\\n LS.rsapss_skey_pre (v modBits) (v eBits) (v dBits) (as_seq h skey) /\\\n bn_v h m < bn_v h (gsub skey 0ul len))\n (ensures fun h0 eq_m h1 -> modifies (loc sgnt) h0 h1 /\\\n (eq_m, as_seq h1 sgnt) == LS.rsapss_sign_compute_sgnt (v modBits) (v eBits) (v dBits) (as_seq h0 skey) (as_seq h0 m))\n\n\ninline_for_extraction noextract\nval rsapss_sign_compute_sgnt:\n #t:limb_t\n -> ke:BE.exp t\n -> modBits:modBits_t t ->\n rsapss_sign_compute_sgnt_st t ke modBits\n\nlet rsapss_sign_compute_sgnt #t ke modBits eBits dBits skey m sgnt =\n push_frame ();\n let h_init = ST.get () in\n [@inline_let] let bits : size_pos = bits t in\n [@inline_let] let numb : size_pos = numbytes t in\n let nLen = blocks modBits (size bits) in\n\n let k = blocks modBits 8ul in\n let s = create nLen (uint #t 0) in\n let m' = create nLen (uint #t 0) in\n let eq_b = rsapss_sign_bn ke modBits eBits dBits skey m m' s in\n LS.blocks_bits_lemma t (v modBits);\n LS.blocks_numb_lemma t (v modBits);\n assert (SD.blocks (v k) numb == v nLen);\n assert (numb * v nLen <= max_size_t);\n BN.bn_to_bytes_be k s sgnt;\n pop_frame ();\n eq_b\n\n\ninline_for_extraction noextract\nlet rsapss_sign_st1 (t:limb_t) (ke:BE.exp t) (a:Hash.hash_alg{S.hash_is_supported a}) (modBits:modBits_t t) =\n let len = blocks modBits (size (bits t)) in\n eBits:size_t\n -> dBits:size_t{LS.skey_len_pre t (v modBits) (v eBits) (v dBits)}\n -> skey:lbignum t (2ul *! len +! blocks eBits (size (bits t)) +! blocks dBits (size (bits t)))\n -> saltLen:size_t\n -> salt:lbuffer uint8 saltLen\n -> msgLen:size_t\n -> msg:lbuffer uint8 msgLen\n -> sgnt:lbuffer uint8 (blocks modBits 8ul) ->\n Stack bool\n (requires fun h -> len == ke.BE.bn.BN.len /\\\n live h salt /\\ live h msg /\\ live h sgnt /\\ live h skey /\\\n disjoint sgnt salt /\\ disjoint sgnt msg /\\ disjoint sgnt salt /\\ disjoint sgnt skey /\\\n disjoint salt msg /\\\n\n LS.rsapss_skey_pre (v modBits) (v eBits) (v dBits) (as_seq h skey) /\\\n LS.rsapss_sign_pre a (v modBits) (v saltLen) (as_seq h salt) (v msgLen) (as_seq h msg))\n (ensures fun h0 eq_m h1 -> modifies (loc sgnt) h0 h1 /\\\n (eq_m, as_seq h1 sgnt) == LS.rsapss_sign_ a (v modBits) (v eBits) (v dBits)\n (as_seq h0 skey) (v saltLen) (as_seq h0 salt) (v msgLen) (as_seq h0 msg))\n\n\ninline_for_extraction noextract\nval rsapss_sign_:\n #t:limb_t\n -> ke:BE.exp t\n -> a:Hash.hash_alg{S.hash_is_supported a}\n -> modBits:modBits_t t ->\n rsapss_sign_st1 t ke a modBits\n\nlet rsapss_sign_ #t ke a modBits eBits dBits skey saltLen salt msgLen msg sgnt =\n push_frame ();\n [@inline_let] let bits : size_pos = bits t in\n let nLen = blocks modBits (size bits) in\n let m = create nLen (uint #t 0) in\n rsapss_sign_msg_to_bn a modBits saltLen salt msgLen msg m;\n let eq_b = rsapss_sign_compute_sgnt ke modBits eBits dBits skey m sgnt in\n pop_frame ();\n eq_b\n\n\ninline_for_extraction noextract\nlet rsapss_sign_st (t:limb_t) (ke:BE.exp t) (a:Hash.hash_alg{S.hash_is_supported a}) (modBits:modBits_t t) =\n let len = blocks modBits (size (bits t)) in\n eBits:size_t\n -> dBits:size_t{LS.skey_len_pre t (v modBits) (v eBits) (v dBits)}\n -> skey:lbignum t (2ul *! len +! blocks eBits (size (bits t)) +! blocks dBits (size (bits t)))\n -> saltLen:size_t\n -> salt:lbuffer uint8 saltLen\n -> msgLen:size_t\n -> msg:lbuffer uint8 msgLen\n -> sgnt:lbuffer uint8 (blocks modBits 8ul) ->\n Stack bool\n (requires fun h -> len == ke.BE.bn.BN.len /\\\n live h salt /\\ live h msg /\\ live h sgnt /\\ live h skey /\\\n disjoint sgnt salt /\\ disjoint sgnt msg /\\ disjoint sgnt salt /\\ disjoint sgnt skey /\\\n disjoint salt msg /\\\n\n LS.rsapss_skey_pre (v modBits) (v eBits) (v dBits) (as_seq h skey))\n (ensures fun h0 b h1 -> modifies (loc sgnt) h0 h1 /\\\n (b, as_seq h1 sgnt) == LS.rsapss_sign a (v modBits) (v eBits) (v dBits)\n (as_seq h0 skey) (v saltLen) (as_seq h0 salt) (v msgLen) (as_seq h0 msg) (as_seq h0 sgnt))\n\n\ninline_for_extraction noextract\nval rsapss_sign:\n #t:limb_t\n -> ke:BE.exp t\n -> a:Hash.hash_alg{S.hash_is_supported a}\n -> modBits:modBits_t t ->\n rsapss_sign_st t ke a modBits\n\nlet rsapss_sign #t ke a modBits eBits dBits skey saltLen salt msgLen msg sgnt =\n let hLen = RM.hash_len a in\n Math.Lemmas.pow2_lt_compat 61 32;\n Math.Lemmas.pow2_lt_compat 125 32;\n //assert (max_size_t < Hash.max_input_length a);\n\n let b =\n saltLen <=. 0xfffffffful -! hLen -! 8ul &&\n saltLen +! hLen +! 2ul <=. blocks (modBits -! 1ul) 8ul in\n\n if b then\n rsapss_sign_ ke a modBits eBits dBits skey saltLen salt msgLen msg sgnt\n else\n false\n\n\n\ninline_for_extraction noextract\nval bn_lt_pow2:\n #t:limb_t\n -> modBits:size_t{1 < v modBits}\n -> m:lbignum t (blocks modBits (size (bits t))) ->\n Stack bool\n (requires fun h -> live h m)\n (ensures fun h0 r h1 -> h0 == h1 /\\\n r == LS.bn_lt_pow2 (v modBits) (as_seq h0 m))\n\nlet bn_lt_pow2 #t modBits m =\n if not ((modBits -! 1ul) %. 8ul =. 0ul) then true\n else begin\n let get_bit = BN.bn_get_ith_bit (blocks modBits (size (bits t))) m (modBits -! 1ul) in\n BB.unsafe_bool_of_limb0 get_bit end\n\n\ninline_for_extraction noextract\nlet rsapss_verify_bn_st (t:limb_t) (ke:BE.exp t) (modBits:modBits_t t) =\n let len = blocks modBits (size (bits t)) in\n eBits:size_t{LS.pkey_len_pre t (v modBits) (v eBits)}\n -> pkey:lbignum t (2ul *! len +! blocks eBits (size (bits t)))\n -> m_def:lbignum t len\n -> s:lbignum t len ->\n Stack bool\n (requires fun h -> len == ke.BE.bn.BN.len /\\\n live h pkey /\\ live h m_def /\\ live h s /\\\n disjoint m_def pkey /\\ disjoint m_def s /\\ disjoint s pkey /\\\n LS.rsapss_pkey_pre (v modBits) (v eBits) (as_seq h pkey))\n (ensures fun h0 r h1 -> modifies (loc m_def) h0 h1 /\\\n (r, as_seq h1 m_def) == LS.rsapss_verify_bn (v modBits) (v eBits) (as_seq h0 pkey) (as_seq h0 m_def) (as_seq h0 s))\n\n\ninline_for_extraction noextract\nval rsapss_verify_bn: #t:limb_t -> ke:BE.exp t -> modBits:modBits_t t -> rsapss_verify_bn_st t ke modBits\nlet rsapss_verify_bn #t ke modBits eBits pkey m_def s =\n [@inline_let] let bits = size (bits t) in\n let nLen = blocks modBits bits in\n let eLen = blocks eBits bits in\n\n let n = sub pkey 0ul nLen in\n let r2 = sub pkey nLen nLen in\n let e = sub pkey (nLen +! nLen) eLen in\n\n let mask = BN.bn_lt_mask nLen s n in\n let h = ST.get () in\n SB.bn_lt_mask_lemma (as_seq h s) (as_seq h n);\n\n let res =\n if BB.unsafe_bool_of_limb mask then begin\n Math.Lemmas.pow2_le_compat (v bits * v nLen) (v modBits);\n SM.bn_precomp_r2_mod_n_lemma (v modBits - 1) (as_seq h n);\n\n let h0 = ST.get () in\n BE.mk_bn_mod_exp_precompr2 nLen ke.BE.exp_vt_precomp n r2 s eBits e m_def;\n let h1 = ST.get () in\n SD.bn_eval_inj (v nLen) (as_seq h1 m_def)\n (SE.bn_mod_exp_vartime_precompr2 (v nLen) (as_seq h0 n) (as_seq h0 r2)\n (as_seq h1 s) (v eBits) (as_seq h0 e));\n\n if bn_lt_pow2 modBits m_def then true\n else false end\n else false in\n res\n\n\ninline_for_extraction noextract\nlet rsapss_verify_bn_to_msg_st (t:limb_t) (a:Hash.hash_alg{S.hash_is_supported a}) (modBits:modBits_t t) =\n saltLen:size_t\n -> msgLen:size_t\n -> msg:lbuffer uint8 msgLen\n -> m:lbignum t (blocks modBits (size (bits t))) ->\n Stack bool\n (requires fun h ->\n live h msg /\\ live h m /\\ disjoint m msg /\\\n\n LS.rsapss_verify_pre a (v saltLen) (v msgLen) (as_seq h msg))\n (ensures fun h0 r h1 -> modifies0 h0 h1 /\\\n r == LS.rsapss_verify_bn_to_msg a (v modBits) (v saltLen) (v msgLen) (as_seq h0 msg) (as_seq h0 m))\n\n\ninline_for_extraction noextract\nval rsapss_verify_bn_to_msg:\n #t:limb_t\n -> a:Hash.hash_alg{S.hash_is_supported a}\n -> modBits:modBits_t t ->\n rsapss_verify_bn_to_msg_st t a modBits\n\nlet rsapss_verify_bn_to_msg #t a modBits saltLen msgLen msg m =\n push_frame ();\n [@inline_let] let bits : size_pos = bits t in\n [@inline_let] let numb : size_pos = numbytes t in\n let nLen = blocks modBits (size bits) in\n\n let emBits = modBits -! 1ul in\n let emLen = blocks emBits 8ul in\n [@inline_let] let mLen = blocks emLen (size numb) in\n let em = create emLen (u8 0) in\n\n LS.blocks_bits_lemma t (v emBits);\n LS.blocks_numb_lemma t (v emBits);\n assert (SD.blocks (v emBits) bits == v mLen);\n assert (numb * v mLen <= max_size_t);\n assert (v mLen <= v nLen);\n\n let m1 = sub m 0ul mLen in\n BN.bn_to_bytes_be emLen m1 em;\n let res = RP.pss_verify a saltLen msgLen msg emBits em in\n pop_frame ();\n res\n\n\ninline_for_extraction noextract\nlet rsapss_verify_compute_msg_st (t:limb_t) (ke:BE.exp t) (modBits:modBits_t t) =\n let len = blocks modBits (size (bits t)) in\n eBits:size_t{LS.pkey_len_pre t (v modBits) (v eBits)}\n -> pkey:lbignum t (2ul *! len +! blocks eBits (size (bits t)))\n -> sgnt:lbuffer uint8 (blocks modBits 8ul)\n -> m:lbignum t len ->\n Stack bool\n (requires fun h -> len == ke.BE.bn.BN.len /\\\n live h sgnt /\\ live h pkey /\\ live h m /\\\n disjoint m sgnt /\\ disjoint m pkey /\\\n as_seq h m == LSeq.create (v len) (uint #t 0) /\\\n LS.rsapss_pkey_pre (v modBits) (v eBits) (as_seq h pkey))\n (ensures fun h0 r h1 -> modifies (loc m) h0 h1 /\\\n (r, as_seq h1 m) == LS.rsapss_verify_compute_msg (v modBits) (v eBits) (as_seq h0 pkey) (as_seq h0 sgnt))\n\n\ninline_for_extraction noextract\nval rsapss_verify_compute_msg:\n #t:limb_t\n -> ke:BE.exp t\n -> modBits:modBits_t t ->\n rsapss_verify_compute_msg_st t ke modBits\n\nlet rsapss_verify_compute_msg #t ke modBits eBits pkey sgnt m =\n push_frame ();\n [@inline_let] let bits : size_pos = bits t in\n [@inline_let] let numb : size_pos = numbytes t in\n let nLen = blocks modBits (size bits) in\n let k = blocks modBits 8ul in\n\n let s = create nLen (uint #t 0) in\n LS.blocks_bits_lemma t (v modBits);\n LS.blocks_numb_lemma t (v modBits);\n assert (SD.blocks (v k) numb == v nLen);\n assert (numb * v nLen <= max_size_t);\n BN.bn_from_bytes_be k sgnt s;\n\n let b = rsapss_verify_bn #t ke modBits eBits pkey m s in\n pop_frame ();\n b\n\n\ninline_for_extraction noextract\nlet rsapss_verify_st1 (t:limb_t) (ke:BE.exp t) (a:Hash.hash_alg{S.hash_is_supported a}) (modBits:modBits_t t) =\n let len = blocks modBits (size (bits t)) in\n eBits:size_t{LS.pkey_len_pre t (v modBits) (v eBits)}\n -> pkey:lbignum t (2ul *! len +! blocks eBits (size (bits t)))\n -> saltLen:size_t\n -> sgnt:lbuffer uint8 (blocks modBits 8ul)\n -> msgLen:size_t\n -> msg:lbuffer uint8 msgLen ->\n Stack bool\n (requires fun h -> len == ke.BE.bn.BN.len /\\\n live h msg /\\ live h sgnt /\\ live h pkey /\\\n disjoint msg sgnt /\\ disjoint msg pkey /\\\n\n LS.rsapss_pkey_pre (v modBits) (v eBits) (as_seq h pkey) /\\\n LS.rsapss_verify_pre a (v saltLen) (v msgLen) (as_seq h msg))\n (ensures fun h0 r h1 -> modifies0 h0 h1 /\\\n r == LS.rsapss_verify_ a (v modBits) (v eBits) (as_seq h0 pkey)\n (v saltLen) (as_seq h0 sgnt) (v msgLen) (as_seq h0 msg))\n\n\ninline_for_extraction noextract\nval rsapss_verify_:\n #t:limb_t\n -> ke:BE.exp t\n -> a:Hash.hash_alg{S.hash_is_supported a}\n -> modBits:modBits_t t ->\n rsapss_verify_st1 t ke a modBits\n\nlet rsapss_verify_ #t ke a modBits eBits pkey saltLen sgnt msgLen msg =\n push_frame ();\n [@inline_let] let bits : size_pos = bits t in\n let nLen = blocks modBits (size bits) in\n let m = create nLen (uint #t 0) in\n let b = rsapss_verify_compute_msg ke modBits eBits pkey sgnt m in\n let res = if b then rsapss_verify_bn_to_msg a modBits saltLen msgLen msg m else false in\n pop_frame ();\n res\n\n\ninline_for_extraction noextract\nlet rsapss_verify_st (t:limb_t) (ke:BE.exp t) (a:Hash.hash_alg{S.hash_is_supported a}) (modBits:modBits_t t) =\n let len = blocks modBits (size (bits t)) in\n eBits:size_t{LS.pkey_len_pre t (v modBits) (v eBits)}\n -> pkey:lbignum t (2ul *! len +! blocks eBits (size (bits t)))\n -> saltLen:size_t\n -> sgntLen:size_t\n -> sgnt:lbuffer uint8 sgntLen\n -> msgLen:size_t\n -> msg:lbuffer uint8 msgLen ->\n Stack bool\n (requires fun h -> len == ke.BE.bn.BN.len /\\\n live h msg /\\ live h sgnt /\\ live h pkey /\\\n disjoint msg sgnt /\\ disjoint msg pkey /\\\n\n LS.rsapss_pkey_pre (v modBits) (v eBits) (as_seq h pkey))\n (ensures fun h0 r h1 -> modifies0 h0 h1 /\\\n r == LS.rsapss_verify a (v modBits) (v eBits) (as_seq h0 pkey)\n (v saltLen) (v sgntLen) (as_seq h0 sgnt) (v msgLen) (as_seq h0 msg))\n\n\ninline_for_extraction noextract\nval rsapss_verify:\n #t:limb_t\n -> ke:BE.exp t\n -> a:Hash.hash_alg{S.hash_is_supported a}\n -> modBits:modBits_t t ->\n rsapss_verify_st t ke a modBits\n\nlet rsapss_verify #t ke a modBits eBits pkey saltLen sgntLen sgnt msgLen msg =\n let hLen = RM.hash_len a in\n Math.Lemmas.pow2_lt_compat 61 32;\n Math.Lemmas.pow2_lt_compat 125 32;\n //assert (max_size_t < Hash.max_input_length a);\n assert (v msgLen <= max_size_t);\n assert (v hLen + 8 < max_size_t);\n\n let b =\n saltLen <=. 0xfffffffful -! hLen -! 8ul &&\n sgntLen =. blocks modBits 8ul in\n\n if b then\n rsapss_verify_ ke a modBits eBits pkey saltLen sgnt msgLen msg\n else\n false\n\n\ninline_for_extraction noextract\nlet rsapss_skey_sign_st (t:limb_t) (ke:BE.exp t) (a:Hash.hash_alg{S.hash_is_supported a}) (modBits:size_t) =\n eBits:size_t\n -> dBits:size_t{LS.skey_len_pre t (v modBits) (v eBits) (v dBits)}\n -> nb:lbuffer uint8 (blocks modBits 8ul)\n -> eb:lbuffer uint8 (blocks eBits 8ul)\n -> db:lbuffer uint8 (blocks dBits 8ul)\n -> saltLen:size_t\n -> salt:lbuffer uint8 saltLen\n -> msgLen:size_t\n -> msg:lbuffer uint8 msgLen\n -> sgnt:lbuffer uint8 (blocks modBits 8ul) ->\n Stack bool\n (requires fun h ->\n blocks modBits (size (bits t)) == ke.BE.bn.BN.len /\\\n live h salt /\\ live h msg /\\ live h sgnt /\\\n live h nb /\\ live h eb /\\ live h db /\\\n disjoint sgnt salt /\\ disjoint sgnt msg /\\ disjoint sgnt salt /\\\n disjoint sgnt nb /\\ disjoint sgnt eb /\\ disjoint sgnt db /\\\n disjoint salt msg)\n (ensures fun h0 b h1 -> modifies (loc sgnt) h0 h1 /\\\n (let sgnt_s = S.rsapss_skey_sign a (v modBits) (v eBits) (v dBits)\n (as_seq h0 nb) (as_seq h0 eb) (as_seq h0 db) (v saltLen) (as_seq h0 salt) (v msgLen) (as_seq h0 msg) in\n if b then Some? sgnt_s /\\ as_seq h1 sgnt == Some?.v sgnt_s else None? sgnt_s))\n\n\ninline_for_extraction noextract\nval rsapss_skey_sign:\n #t:limb_t\n -> ke:BE.exp t\n -> a:Hash.hash_alg{S.hash_is_supported a}\n -> modBits:modBits_t t\n -> rsapss_load_skey:RK.rsapss_load_skey_st t ke modBits\n -> rsapss_sign:rsapss_sign_st t ke a modBits ->\n rsapss_skey_sign_st t ke a modBits\n\nlet rsapss_skey_sign #t ke a modBits rsapss_load_skey rsapss_sign eBits dBits nb eb db saltLen salt msgLen msg sgnt =\n [@inline_let] let bits = size (bits t) in\n let h0 = ST.get () in\n push_frame ();\n let skey = create (2ul *! blocks modBits bits +! blocks eBits bits +! blocks dBits bits) (uint #t 0) in\n let b = rsapss_load_skey eBits dBits nb eb db skey in\n LS.rsapss_load_skey_lemma #t (v modBits) (v eBits) (v dBits) (as_seq h0 nb) (as_seq h0 eb) (as_seq h0 db);\n\n let res =\n if b then\n rsapss_sign eBits dBits skey saltLen salt msgLen msg sgnt\n else\n false in\n pop_frame ();\n let h1 = ST.get () in\n assert ((res, as_seq h1 sgnt) == LS.rsapss_skey_sign #t a (v modBits) (v eBits) (v dBits)\n (as_seq h0 nb) (as_seq h0 eb) (as_seq h0 db) (v saltLen) (as_seq h0 salt)\n (v msgLen) (as_seq h0 msg) (as_seq h0 sgnt));\n res\n\n\ninline_for_extraction noextract\nlet rsapss_pkey_verify_st (t:limb_t) (ke:BE.exp t) (a:Hash.hash_alg{S.hash_is_supported a}) (modBits:size_t) =\n eBits:size_t{LS.pkey_len_pre t (v modBits) (v eBits)}\n -> nb:lbuffer uint8 (blocks modBits 8ul)\n -> eb:lbuffer uint8 (blocks eBits 8ul)\n -> saltLen:size_t\n -> sgntLen:size_t\n -> sgnt:lbuffer uint8 sgntLen\n -> msgLen:size_t\n -> msg:lbuffer uint8 msgLen ->\n Stack bool\n (requires fun h ->\n blocks modBits (size (bits t)) == ke.BE.bn.BN.len /\\\n live h msg /\\ live h sgnt /\\ live h nb /\\ live h eb /\\\n disjoint msg sgnt /\\ disjoint nb eb /\\ disjoint sgnt nb /\\\n disjoint sgnt eb /\\ disjoint msg nb /\\ disjoint msg eb)\n (ensures fun h0 r h1 -> modifies0 h0 h1 /\\\n r == S.rsapss_pkey_verify a (v modBits) (v eBits) (as_seq h0 nb) (as_seq h0 eb)\n (v saltLen) (v sgntLen) (as_seq h0 sgnt) (v msgLen) (as_seq h0 msg))\n\n\ninline_for_extraction noextract\nval rsapss_pkey_verify:\n #t:limb_t\n -> ke:BE.exp t\n -> a:Hash.hash_alg{S.hash_is_supported a}\n -> modBits:modBits_t t\n -> rsapss_load_pkey:RK.rsapss_load_pkey_st t ke modBits\n -> rsapss_verify:rsapss_verify_st t ke a modBits ->\n rsapss_pkey_verify_st t ke a modBits\n\nPartial implementation:\nlet rsapss_pkey_verify\n #t\n ke\n a\n modBits\n rsapss_load_pkey\n rsapss_verify\n eBits\n nb\n eb\n saltLen\n sgntLen\n sgnt\n msgLen\n msg\n =\n \n"
},
{
"role": "assistant",
"content": "push_frame ();\n[@@ inline_let ]let bits = size (bits t) in\nlet pkey = create (2ul *! blocks modBits bits +! blocks eBits bits) (uint #t 0) in\nlet h0 = ST.get () in\nlet b = rsapss_load_pkey eBits nb eb pkey in\nLS.rsapss_load_pkey_lemma #t (v modBits) (v eBits) (as_seq h0 nb) (as_seq h0 eb);\nlet res = if b then rsapss_verify eBits pkey saltLen sgntLen sgnt msgLen msg else false in\npop_frame ();\nlet h1 = ST.get () in\nassert (res ==\n LS.rsapss_pkey_verify #t a (v modBits) (v eBits) (as_seq h0 nb) (as_seq h0 eb) (v saltLen)\n (v sgntLen) (as_seq h0 sgnt) (v msgLen) (as_seq h0 msg));\nres"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n h: FStar.Monotonic.HyperStack.mem ->\n c: LowStar.Lib.LinkedList.t a ->\n l1: Prims.list a ->\n l2: Prims.list a\n -> FStar.Pervasives.Lemma\n (requires\n LowStar.Lib.LinkedList.well_formed h c l1 /\\ LowStar.Lib.LinkedList.well_formed h c l2)\n (ensures l1 == l2)\n (decreases FStar.Ghost.reveal (FStar.Ghost.hide l1))\n [\n SMTPat (LowStar.Lib.LinkedList.well_formed h c l1);\n SMTPat (LowStar.Lib.LinkedList.well_formed h c l2)\n ]\nEffect: FStar.Pervasives.Lemma\n\nContext:\nmodule LowStar.Lib.LinkedList\n\n/// This module, oftentimes referred to as LL1 in other parts of the\n/// documentation, provides a low-level view over a linked list. Unless you\n/// intend on modifying the structure of the linked list (e.g. removing cells)\n/// or iterating over each cell, your are probably better off using LL2.\n///\n/// This module is intentionally not relying on a tight abstraction, to i) permit\n/// direct-style iteration by clients and ii) allow stating some SMTPat lemmas\n/// only on LL1 rather than replicating them in LL2 (going from LL2 to LL1 can\n/// happen automatically at the SMT level since LL2 does not have an fsti either).\n///\n/// Some open questions relative to the design of this module:\n/// - should ``v`` be defined directly in terms of ``cells`` rather than have a\n/// recursive definition along with an equivalence lemma?\n/// - should ``footprint`` be defined directly in terms of ``cells`` rather than\n/// have a recursive definition along with an equivalence lemma?\n/// - is it worth the trouble defining a footprint? my intuition is that it's\n/// fine and allows some clients to do very precise reasoning -- other modules\n/// (e.g. AssocList), when establishing abstraction boundaries, can simplify\n/// things and get rid of the precise footprint reasoning in favor of a static\n/// region-based footprint\n///\n/// Some remaining work items:\n/// - switch to C.Loops.while\n\nopen LowStar.BufferOps\n\nmodule B = LowStar.Buffer\nmodule HS = FStar.HyperStack\nmodule G = FStar.Ghost\nmodule L = FStar.List.Tot\nmodule U32 = FStar.UInt32\nmodule ST = FStar.HyperStack.ST\n\nopen FStar.HyperStack.ST\n\n/// Breaking from my own recommendation here... using fuel 1 / ifuel 1 since\n/// everything does recursive list reasoning.\n#set-options \"--__no_positivity --fuel 1 --ifuel 1\"\n\n/// Definition of the Low* type\n/// ---------------------------\nnoeq\ntype t (a: Type0) =\n B.pointer_or_null (cell a)\n\nand cell (a: Type0) = {\n next: t a;\n data: a;\n}\n\n/// Unlike the canonical style that associates a ``v`` function to any stateful\n/// representation, we use a relation here. Writing a function is just\n/// impossible, since we can't prove termination of ``v`` owing to the fact that\n/// there may be an infinite number of cells in the heap.\n///\n/// LL2 is the module that switches to the function style.\n///\n/// Note: no need to use erased here for ``l`` since the function is already in ``GTot``.\nlet rec well_formed #a (h: HS.mem) (c: t a) (l: list a): GTot Type0 (decreases l) =\n B.live h c /\\ (\n match l with\n | [] ->\n B.g_is_null c\n | a :: q ->\n B.length c == 1 /\\ (\n let { data=data; next=next } = B.get h c 0 in\n a == data /\\\n well_formed h next (G.hide q)\n ))\n\n/// Precise reasoning via the cells predicate\n/// -----------------------------------------\n///\n/// The modifies clauses are given either on the footprint of the list, or on a\n/// region that includes the footprint of the list (higher up the stack, e.g.\n/// LL2). This is good, but does not rule out \"stupid\" implementations that\n/// might re-allocate every single list cell when, say, doing a pop.\n///\n/// For advanced usages which take pointers directly to individual list cells,\n/// it's important to rule out these cases. We thus define the ``cells``\n/// predicate for that purpose.\n\nlet rec cells #a (h: HS.mem) (c: t a) (l: list a): Ghost (list (B.pointer (cell a)))\n (requires well_formed h c l)\n (ensures fun _ -> True)\n (decreases l)\n=\n if B.g_is_null c then\n []\n else\n c :: cells h (B.deref h c).next (List.Tot.tl l)\n\n/// A missing bit of automated reasoning\n/// ------------------------------------\n\n/// This should be absolutely trivial, yet, because there's no pattern on\n/// null_unique, we have to do the case analysis and the decomposition\n/// ourselves. This is quite tedious, as it usually takes a while (at least it\n/// did for me!) to realize that this property does not hold automatically.\n///\n/// I'm setting an SMTPat on this lemma because it's absolutely impossible to\n/// get any work done without it.\nlet same_cells_same_pointer #a (h0 h1: HS.mem) (ll0 ll1: t a) (l0 l1: list a): Lemma\n (requires (\n well_formed h0 ll0 l0 /\\\n well_formed h1 ll1 l1 /\\\n cells h0 ll0 l0 == cells h1 ll1 l1))\n (ensures (\n ll0 == ll1))\n [ SMTPat (cells h0 ll0 l0); SMTPat (cells h1 ll1 l1) ]\n=\n if B.g_is_null ll0 && B.g_is_null ll1 then begin\n B.null_unique ll0;\n B.null_unique ll1\n end else if not (B.g_is_null ll0) && not (B.g_is_null ll1) then begin\n ()\n end else\n false_elim ()\n\n/// Classic stateful reasoning lemmas\n/// ---------------------------------\n///\n/// This essentially follows ``stateful``, the type class of a mutable piece of\n/// state defined using the modifies-theory of Low*v2 and equipped with the\n/// right lemmas to enable usage by clients. See\n/// hacl-star/code/streaming/Hacl.Streaming.Interface.fst for more context.\n\nval footprint: (#a: Type) -> (h: HS.mem) -> (l: t a) -> (n: list a) -> Ghost B.loc\n (requires (well_formed h l n))\n (ensures (fun refs -> True))\n (decreases n)\n\nlet rec footprint #a h l n =\n if B.g_is_null l\n then B.loc_none\n else\n let {next = next} = B.get h l 0 in\n let refs = footprint h next (G.hide (L.tl n)) in\n B.loc_union (B.loc_addr_of_buffer l) refs\n\n/// Departing from LinkedList4 here. I prefer to bolt these into the invariant\n/// rather than requiring on some wizard lemmas from LowStar.Monotonic.Buffer to\n/// deduce these.\nval cells_pairwise_disjoint: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_pairwise_disjoint #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.loc_disjoint (B.loc_addr_of_buffer l) (footprint h next (L.tl n)) /\\\n cells_pairwise_disjoint h next (L.tl n)\n\nval cells_live_freeable: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_live_freeable #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.live h l /\\ B.freeable l /\\\n cells_live_freeable h next (L.tl n)\n\nval invariant: #a:Type -> h:HS.mem -> l: t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\nlet invariant #a h l n =\n cells_live_freeable h l n /\\ cells_pairwise_disjoint h l n\n\n/// Normally this would be automatic and writing a custom lemma for that stuff\n/// would be needed only when writing an fsti, however, as stated earlier,\n/// recursion makes everything more difficult and this absolutely needs to be in\n/// scope otherwise the modifies-clause theory cannot deduce disjointness of\n/// fresh allocations w.r.t. the footprint.\nlet rec invariant_loc_in_footprint (#a: Type) (h: HS.mem) (l: t a) (n: list a): Lemma\n (requires well_formed h l n)\n (ensures B.loc_in (footprint h l n) h)\n (decreases n)\n [ SMTPat (invariant h l n) ]\n=\n if B.g_is_null l then\n ()\n else\n let next = (B.deref h l).next in\n invariant_loc_in_footprint h next (L.tl n)\n\n/// Another absolutely essential lemma, for which interestingly the pattern\n/// \"invariant h1 l n\" does not work, but well_formed does... I guess that's\n/// alright.\nlet rec frame (#a: Type) (l: t a) (n: list a) (r: B.loc) (h0 h1: HS.mem): Lemma\n (requires (\n well_formed h0 l n /\\\n invariant h0 l n /\\\n B.loc_disjoint r (footprint h0 l n) /\\\n B.modifies r h0 h1\n ))\n (ensures (\n well_formed h1 l n /\\\n footprint h1 l n == footprint h0 l n /\\\n cells h1 l n == cells h0 l n /\\\n invariant h1 l n\n ))\n (decreases n)\n [ SMTPat (well_formed h1 l n); SMTPat (B.modifies r h0 h1) ]\n=\n if B.g_is_null l then\n ()\n else\n frame (B.deref h0 l).next (L.tl n) r h0 h1\n\n/// Lemmas for working with linked lists\n/// ------------------------------------\n///\n/// Since it's a recursive data structure, a lot of the automated reasoning from\n/// LowStar.Buffer doesn't work since it requires inductions. Some helpful\n/// properties are therefore proved here.\n\nPartial implementation:\nlet rec length_functional #a (h: HS.mem) (c: t a) (l1: list a) (l2: list a)\n : Lemma (requires (well_formed h c l1 /\\ well_formed h c l2))\n (ensures (l1 == l2))\n (decreases (G.reveal l1))\n [SMTPat (well_formed h c l1); SMTPat (well_formed h c l2)] =\n \n"
},
{
"role": "assistant",
"content": "if B.g_is_null c\nthen ()\nelse\n let { next = next } = B.get h c 0 in\n length_functional h next (G.hide (L.tl l1)) (G.hide (L.tl l2))"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: h: FStar.Monotonic.HyperStack.mem -> l: LowStar.Lib.LinkedList.t a -> n: Prims.list a\n -> FStar.Pervasives.Lemma (requires LowStar.Lib.LinkedList.well_formed h l n)\n (ensures LowStar.Monotonic.Buffer.loc_in (LowStar.Lib.LinkedList.footprint h l n) h)\n (decreases n)\n [SMTPat (LowStar.Lib.LinkedList.invariant h l n)]\nEffect: FStar.Pervasives.Lemma\n\nContext:\nmodule LowStar.Lib.LinkedList\n\n/// This module, oftentimes referred to as LL1 in other parts of the\n/// documentation, provides a low-level view over a linked list. Unless you\n/// intend on modifying the structure of the linked list (e.g. removing cells)\n/// or iterating over each cell, your are probably better off using LL2.\n///\n/// This module is intentionally not relying on a tight abstraction, to i) permit\n/// direct-style iteration by clients and ii) allow stating some SMTPat lemmas\n/// only on LL1 rather than replicating them in LL2 (going from LL2 to LL1 can\n/// happen automatically at the SMT level since LL2 does not have an fsti either).\n///\n/// Some open questions relative to the design of this module:\n/// - should ``v`` be defined directly in terms of ``cells`` rather than have a\n/// recursive definition along with an equivalence lemma?\n/// - should ``footprint`` be defined directly in terms of ``cells`` rather than\n/// have a recursive definition along with an equivalence lemma?\n/// - is it worth the trouble defining a footprint? my intuition is that it's\n/// fine and allows some clients to do very precise reasoning -- other modules\n/// (e.g. AssocList), when establishing abstraction boundaries, can simplify\n/// things and get rid of the precise footprint reasoning in favor of a static\n/// region-based footprint\n///\n/// Some remaining work items:\n/// - switch to C.Loops.while\n\nopen LowStar.BufferOps\n\nmodule B = LowStar.Buffer\nmodule HS = FStar.HyperStack\nmodule G = FStar.Ghost\nmodule L = FStar.List.Tot\nmodule U32 = FStar.UInt32\nmodule ST = FStar.HyperStack.ST\n\nopen FStar.HyperStack.ST\n\n/// Breaking from my own recommendation here... using fuel 1 / ifuel 1 since\n/// everything does recursive list reasoning.\n#set-options \"--__no_positivity --fuel 1 --ifuel 1\"\n\n/// Definition of the Low* type\n/// ---------------------------\nnoeq\ntype t (a: Type0) =\n B.pointer_or_null (cell a)\n\nand cell (a: Type0) = {\n next: t a;\n data: a;\n}\n\n/// Unlike the canonical style that associates a ``v`` function to any stateful\n/// representation, we use a relation here. Writing a function is just\n/// impossible, since we can't prove termination of ``v`` owing to the fact that\n/// there may be an infinite number of cells in the heap.\n///\n/// LL2 is the module that switches to the function style.\n///\n/// Note: no need to use erased here for ``l`` since the function is already in ``GTot``.\nlet rec well_formed #a (h: HS.mem) (c: t a) (l: list a): GTot Type0 (decreases l) =\n B.live h c /\\ (\n match l with\n | [] ->\n B.g_is_null c\n | a :: q ->\n B.length c == 1 /\\ (\n let { data=data; next=next } = B.get h c 0 in\n a == data /\\\n well_formed h next (G.hide q)\n ))\n\n/// Precise reasoning via the cells predicate\n/// -----------------------------------------\n///\n/// The modifies clauses are given either on the footprint of the list, or on a\n/// region that includes the footprint of the list (higher up the stack, e.g.\n/// LL2). This is good, but does not rule out \"stupid\" implementations that\n/// might re-allocate every single list cell when, say, doing a pop.\n///\n/// For advanced usages which take pointers directly to individual list cells,\n/// it's important to rule out these cases. We thus define the ``cells``\n/// predicate for that purpose.\n\nlet rec cells #a (h: HS.mem) (c: t a) (l: list a): Ghost (list (B.pointer (cell a)))\n (requires well_formed h c l)\n (ensures fun _ -> True)\n (decreases l)\n=\n if B.g_is_null c then\n []\n else\n c :: cells h (B.deref h c).next (List.Tot.tl l)\n\n/// A missing bit of automated reasoning\n/// ------------------------------------\n\n/// This should be absolutely trivial, yet, because there's no pattern on\n/// null_unique, we have to do the case analysis and the decomposition\n/// ourselves. This is quite tedious, as it usually takes a while (at least it\n/// did for me!) to realize that this property does not hold automatically.\n///\n/// I'm setting an SMTPat on this lemma because it's absolutely impossible to\n/// get any work done without it.\nlet same_cells_same_pointer #a (h0 h1: HS.mem) (ll0 ll1: t a) (l0 l1: list a): Lemma\n (requires (\n well_formed h0 ll0 l0 /\\\n well_formed h1 ll1 l1 /\\\n cells h0 ll0 l0 == cells h1 ll1 l1))\n (ensures (\n ll0 == ll1))\n [ SMTPat (cells h0 ll0 l0); SMTPat (cells h1 ll1 l1) ]\n=\n if B.g_is_null ll0 && B.g_is_null ll1 then begin\n B.null_unique ll0;\n B.null_unique ll1\n end else if not (B.g_is_null ll0) && not (B.g_is_null ll1) then begin\n ()\n end else\n false_elim ()\n\n/// Classic stateful reasoning lemmas\n/// ---------------------------------\n///\n/// This essentially follows ``stateful``, the type class of a mutable piece of\n/// state defined using the modifies-theory of Low*v2 and equipped with the\n/// right lemmas to enable usage by clients. See\n/// hacl-star/code/streaming/Hacl.Streaming.Interface.fst for more context.\n\nval footprint: (#a: Type) -> (h: HS.mem) -> (l: t a) -> (n: list a) -> Ghost B.loc\n (requires (well_formed h l n))\n (ensures (fun refs -> True))\n (decreases n)\n\nlet rec footprint #a h l n =\n if B.g_is_null l\n then B.loc_none\n else\n let {next = next} = B.get h l 0 in\n let refs = footprint h next (G.hide (L.tl n)) in\n B.loc_union (B.loc_addr_of_buffer l) refs\n\n/// Departing from LinkedList4 here. I prefer to bolt these into the invariant\n/// rather than requiring on some wizard lemmas from LowStar.Monotonic.Buffer to\n/// deduce these.\nval cells_pairwise_disjoint: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_pairwise_disjoint #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.loc_disjoint (B.loc_addr_of_buffer l) (footprint h next (L.tl n)) /\\\n cells_pairwise_disjoint h next (L.tl n)\n\nval cells_live_freeable: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_live_freeable #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.live h l /\\ B.freeable l /\\\n cells_live_freeable h next (L.tl n)\n\nval invariant: #a:Type -> h:HS.mem -> l: t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\nlet invariant #a h l n =\n cells_live_freeable h l n /\\ cells_pairwise_disjoint h l n\n\n/// Normally this would be automatic and writing a custom lemma for that stuff\n/// would be needed only when writing an fsti, however, as stated earlier,\n/// recursion makes everything more difficult and this absolutely needs to be in\n/// scope otherwise the modifies-clause theory cannot deduce disjointness of\n\nPartial implementation:\nlet rec invariant_loc_in_footprint (#a: Type) (h: HS.mem) (l: t a) (n: list a)\n : Lemma (requires well_formed h l n)\n (ensures B.loc_in (footprint h l n) h)\n (decreases n)\n [SMTPat (invariant h l n)] =\n \n"
},
{
"role": "assistant",
"content": "if B.g_is_null l\nthen ()\nelse\n let next = (B.deref h l).next in\n invariant_loc_in_footprint h next (L.tl n)"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: h: FStar.Monotonic.HyperStack.mem -> ll: LowStar.Lib.LinkedList.t a -> l: Prims.list a\n -> FStar.Pervasives.Lemma\n (requires LowStar.Lib.LinkedList.well_formed h ll l /\\ LowStar.Lib.LinkedList.invariant h ll l\n )\n (ensures\n LowStar.Lib.LinkedList.gmap (LowStar.Lib.LinkedList.deref_data h)\n (LowStar.Lib.LinkedList.cells h ll l) ==\n l)\n (decreases l)\n [SMTPat (LowStar.Lib.LinkedList.well_formed h ll l)]\nEffect: FStar.Pervasives.Lemma\n\nContext:\nmodule LowStar.Lib.LinkedList\n\n/// This module, oftentimes referred to as LL1 in other parts of the\n/// documentation, provides a low-level view over a linked list. Unless you\n/// intend on modifying the structure of the linked list (e.g. removing cells)\n/// or iterating over each cell, your are probably better off using LL2.\n///\n/// This module is intentionally not relying on a tight abstraction, to i) permit\n/// direct-style iteration by clients and ii) allow stating some SMTPat lemmas\n/// only on LL1 rather than replicating them in LL2 (going from LL2 to LL1 can\n/// happen automatically at the SMT level since LL2 does not have an fsti either).\n///\n/// Some open questions relative to the design of this module:\n/// - should ``v`` be defined directly in terms of ``cells`` rather than have a\n/// recursive definition along with an equivalence lemma?\n/// - should ``footprint`` be defined directly in terms of ``cells`` rather than\n/// have a recursive definition along with an equivalence lemma?\n/// - is it worth the trouble defining a footprint? my intuition is that it's\n/// fine and allows some clients to do very precise reasoning -- other modules\n/// (e.g. AssocList), when establishing abstraction boundaries, can simplify\n/// things and get rid of the precise footprint reasoning in favor of a static\n/// region-based footprint\n///\n/// Some remaining work items:\n/// - switch to C.Loops.while\n\nopen LowStar.BufferOps\n\nmodule B = LowStar.Buffer\nmodule HS = FStar.HyperStack\nmodule G = FStar.Ghost\nmodule L = FStar.List.Tot\nmodule U32 = FStar.UInt32\nmodule ST = FStar.HyperStack.ST\n\nopen FStar.HyperStack.ST\n\n/// Breaking from my own recommendation here... using fuel 1 / ifuel 1 since\n/// everything does recursive list reasoning.\n#set-options \"--__no_positivity --fuel 1 --ifuel 1\"\n\n/// Definition of the Low* type\n/// ---------------------------\nnoeq\ntype t (a: Type0) =\n B.pointer_or_null (cell a)\n\nand cell (a: Type0) = {\n next: t a;\n data: a;\n}\n\n/// Unlike the canonical style that associates a ``v`` function to any stateful\n/// representation, we use a relation here. Writing a function is just\n/// impossible, since we can't prove termination of ``v`` owing to the fact that\n/// there may be an infinite number of cells in the heap.\n///\n/// LL2 is the module that switches to the function style.\n///\n/// Note: no need to use erased here for ``l`` since the function is already in ``GTot``.\nlet rec well_formed #a (h: HS.mem) (c: t a) (l: list a): GTot Type0 (decreases l) =\n B.live h c /\\ (\n match l with\n | [] ->\n B.g_is_null c\n | a :: q ->\n B.length c == 1 /\\ (\n let { data=data; next=next } = B.get h c 0 in\n a == data /\\\n well_formed h next (G.hide q)\n ))\n\n/// Precise reasoning via the cells predicate\n/// -----------------------------------------\n///\n/// The modifies clauses are given either on the footprint of the list, or on a\n/// region that includes the footprint of the list (higher up the stack, e.g.\n/// LL2). This is good, but does not rule out \"stupid\" implementations that\n/// might re-allocate every single list cell when, say, doing a pop.\n///\n/// For advanced usages which take pointers directly to individual list cells,\n/// it's important to rule out these cases. We thus define the ``cells``\n/// predicate for that purpose.\n\nlet rec cells #a (h: HS.mem) (c: t a) (l: list a): Ghost (list (B.pointer (cell a)))\n (requires well_formed h c l)\n (ensures fun _ -> True)\n (decreases l)\n=\n if B.g_is_null c then\n []\n else\n c :: cells h (B.deref h c).next (List.Tot.tl l)\n\n/// A missing bit of automated reasoning\n/// ------------------------------------\n\n/// This should be absolutely trivial, yet, because there's no pattern on\n/// null_unique, we have to do the case analysis and the decomposition\n/// ourselves. This is quite tedious, as it usually takes a while (at least it\n/// did for me!) to realize that this property does not hold automatically.\n///\n/// I'm setting an SMTPat on this lemma because it's absolutely impossible to\n/// get any work done without it.\nlet same_cells_same_pointer #a (h0 h1: HS.mem) (ll0 ll1: t a) (l0 l1: list a): Lemma\n (requires (\n well_formed h0 ll0 l0 /\\\n well_formed h1 ll1 l1 /\\\n cells h0 ll0 l0 == cells h1 ll1 l1))\n (ensures (\n ll0 == ll1))\n [ SMTPat (cells h0 ll0 l0); SMTPat (cells h1 ll1 l1) ]\n=\n if B.g_is_null ll0 && B.g_is_null ll1 then begin\n B.null_unique ll0;\n B.null_unique ll1\n end else if not (B.g_is_null ll0) && not (B.g_is_null ll1) then begin\n ()\n end else\n false_elim ()\n\n/// Classic stateful reasoning lemmas\n/// ---------------------------------\n///\n/// This essentially follows ``stateful``, the type class of a mutable piece of\n/// state defined using the modifies-theory of Low*v2 and equipped with the\n/// right lemmas to enable usage by clients. See\n/// hacl-star/code/streaming/Hacl.Streaming.Interface.fst for more context.\n\nval footprint: (#a: Type) -> (h: HS.mem) -> (l: t a) -> (n: list a) -> Ghost B.loc\n (requires (well_formed h l n))\n (ensures (fun refs -> True))\n (decreases n)\n\nlet rec footprint #a h l n =\n if B.g_is_null l\n then B.loc_none\n else\n let {next = next} = B.get h l 0 in\n let refs = footprint h next (G.hide (L.tl n)) in\n B.loc_union (B.loc_addr_of_buffer l) refs\n\n/// Departing from LinkedList4 here. I prefer to bolt these into the invariant\n/// rather than requiring on some wizard lemmas from LowStar.Monotonic.Buffer to\n/// deduce these.\nval cells_pairwise_disjoint: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_pairwise_disjoint #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.loc_disjoint (B.loc_addr_of_buffer l) (footprint h next (L.tl n)) /\\\n cells_pairwise_disjoint h next (L.tl n)\n\nval cells_live_freeable: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_live_freeable #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.live h l /\\ B.freeable l /\\\n cells_live_freeable h next (L.tl n)\n\nval invariant: #a:Type -> h:HS.mem -> l: t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\nlet invariant #a h l n =\n cells_live_freeable h l n /\\ cells_pairwise_disjoint h l n\n\n/// Normally this would be automatic and writing a custom lemma for that stuff\n/// would be needed only when writing an fsti, however, as stated earlier,\n/// recursion makes everything more difficult and this absolutely needs to be in\n/// scope otherwise the modifies-clause theory cannot deduce disjointness of\n/// fresh allocations w.r.t. the footprint.\nlet rec invariant_loc_in_footprint (#a: Type) (h: HS.mem) (l: t a) (n: list a): Lemma\n (requires well_formed h l n)\n (ensures B.loc_in (footprint h l n) h)\n (decreases n)\n [ SMTPat (invariant h l n) ]\n=\n if B.g_is_null l then\n ()\n else\n let next = (B.deref h l).next in\n invariant_loc_in_footprint h next (L.tl n)\n\n/// Another absolutely essential lemma, for which interestingly the pattern\n/// \"invariant h1 l n\" does not work, but well_formed does... I guess that's\n/// alright.\nlet rec frame (#a: Type) (l: t a) (n: list a) (r: B.loc) (h0 h1: HS.mem): Lemma\n (requires (\n well_formed h0 l n /\\\n invariant h0 l n /\\\n B.loc_disjoint r (footprint h0 l n) /\\\n B.modifies r h0 h1\n ))\n (ensures (\n well_formed h1 l n /\\\n footprint h1 l n == footprint h0 l n /\\\n cells h1 l n == cells h0 l n /\\\n invariant h1 l n\n ))\n (decreases n)\n [ SMTPat (well_formed h1 l n); SMTPat (B.modifies r h0 h1) ]\n=\n if B.g_is_null l then\n ()\n else\n frame (B.deref h0 l).next (L.tl n) r h0 h1\n\n/// Lemmas for working with linked lists\n/// ------------------------------------\n///\n/// Since it's a recursive data structure, a lot of the automated reasoning from\n/// LowStar.Buffer doesn't work since it requires inductions. Some helpful\n/// properties are therefore proved here.\n\nlet rec length_functional #a (h: HS.mem) (c: t a) (l1 l2: list a):\n Lemma\n (requires (well_formed h c l1 /\\ well_formed h c l2))\n (ensures (l1 == l2))\n (decreases (G.reveal l1))\n [ SMTPat (well_formed h c l1); SMTPat (well_formed h c l2) ] =\n if B.g_is_null c\n then ()\n else\n let { next=next } = B.get h c 0 in\n length_functional h next (G.hide (L.tl l1)) (G.hide (L.tl l2))\n\n/// The footprint is based on loc_addr_of_buffer so that we can write a free\n/// operation that operates on the footprint. However, this invalidates most of\n/// the helper lemmas for disjointness that were present in the previous\n/// iteration of this module, named LinkedList4.fst. It's ok, we bake\n/// disjointness directly onto the invariant and it works just as well.\n\n/// Connection between ``cells`` and ``v``\n/// --------------------------------------\n\n/// Redefining some helpers.\n#push-options \"--ifuel 1\"\nlet rec gmap #a #b (f: a -> GTot b) (xs: list a): GTot (list b) =\n match xs with\n | [] -> []\n | x :: xs -> f x :: gmap f xs\n\n/// \"As we all know\", right folds are the easiest to work with because they\n/// follow the natural structure of recursion (left-folds are evil).\nlet rec gfold_right #a #b (f: b -> a -> GTot b) (xs: list a) (acc: b): Ghost b\n (requires True)\n (ensures fun _ -> True)\n (decreases xs)\n=\n match xs with\n | [] -> acc\n | x :: xs -> f (gfold_right f xs acc) x\n#pop-options\n\n/// Connecting ``v`` and ``cells``.\n///\n/// See ``tests/Wireguard.fst`` for why going through gmap is important and how\n/// to structure your definitions (should you need to reason about ``cells``) in\n/// a way that will make your life easier.\nlet deref_data #a (h: HS.mem) (c: B.pointer (cell a)): GTot a =\n (B.deref h c).data\n\nPartial implementation:\nlet rec deref_cells_is_v #a (h: HS.mem) (ll: t a) (l: list a)\n : Lemma (requires well_formed h ll l /\\ invariant h ll l)\n (ensures gmap (deref_data h) (cells h ll l) == l)\n (decreases l)\n [SMTPat (well_formed h ll l)] =\n \n"
},
{
"role": "assistant",
"content": "if B.g_is_null ll then () else deref_cells_is_v h (B.deref h ll).next (List.Tot.tl l)"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: l: LowStar.Lib.LinkedList.t a -> FStar.HyperStack.ST.ST Prims.unit\nEffect: FStar.HyperStack.ST.ST\n\nContext:\nmodule LowStar.Lib.LinkedList\n\n/// This module, oftentimes referred to as LL1 in other parts of the\n/// documentation, provides a low-level view over a linked list. Unless you\n/// intend on modifying the structure of the linked list (e.g. removing cells)\n/// or iterating over each cell, your are probably better off using LL2.\n///\n/// This module is intentionally not relying on a tight abstraction, to i) permit\n/// direct-style iteration by clients and ii) allow stating some SMTPat lemmas\n/// only on LL1 rather than replicating them in LL2 (going from LL2 to LL1 can\n/// happen automatically at the SMT level since LL2 does not have an fsti either).\n///\n/// Some open questions relative to the design of this module:\n/// - should ``v`` be defined directly in terms of ``cells`` rather than have a\n/// recursive definition along with an equivalence lemma?\n/// - should ``footprint`` be defined directly in terms of ``cells`` rather than\n/// have a recursive definition along with an equivalence lemma?\n/// - is it worth the trouble defining a footprint? my intuition is that it's\n/// fine and allows some clients to do very precise reasoning -- other modules\n/// (e.g. AssocList), when establishing abstraction boundaries, can simplify\n/// things and get rid of the precise footprint reasoning in favor of a static\n/// region-based footprint\n///\n/// Some remaining work items:\n/// - switch to C.Loops.while\n\nopen LowStar.BufferOps\n\nmodule B = LowStar.Buffer\nmodule HS = FStar.HyperStack\nmodule G = FStar.Ghost\nmodule L = FStar.List.Tot\nmodule U32 = FStar.UInt32\nmodule ST = FStar.HyperStack.ST\n\nopen FStar.HyperStack.ST\n\n/// Breaking from my own recommendation here... using fuel 1 / ifuel 1 since\n/// everything does recursive list reasoning.\n#set-options \"--__no_positivity --fuel 1 --ifuel 1\"\n\n/// Definition of the Low* type\n/// ---------------------------\nnoeq\ntype t (a: Type0) =\n B.pointer_or_null (cell a)\n\nand cell (a: Type0) = {\n next: t a;\n data: a;\n}\n\n/// Unlike the canonical style that associates a ``v`` function to any stateful\n/// representation, we use a relation here. Writing a function is just\n/// impossible, since we can't prove termination of ``v`` owing to the fact that\n/// there may be an infinite number of cells in the heap.\n///\n/// LL2 is the module that switches to the function style.\n///\n/// Note: no need to use erased here for ``l`` since the function is already in ``GTot``.\nlet rec well_formed #a (h: HS.mem) (c: t a) (l: list a): GTot Type0 (decreases l) =\n B.live h c /\\ (\n match l with\n | [] ->\n B.g_is_null c\n | a :: q ->\n B.length c == 1 /\\ (\n let { data=data; next=next } = B.get h c 0 in\n a == data /\\\n well_formed h next (G.hide q)\n ))\n\n/// Precise reasoning via the cells predicate\n/// -----------------------------------------\n///\n/// The modifies clauses are given either on the footprint of the list, or on a\n/// region that includes the footprint of the list (higher up the stack, e.g.\n/// LL2). This is good, but does not rule out \"stupid\" implementations that\n/// might re-allocate every single list cell when, say, doing a pop.\n///\n/// For advanced usages which take pointers directly to individual list cells,\n/// it's important to rule out these cases. We thus define the ``cells``\n/// predicate for that purpose.\n\nlet rec cells #a (h: HS.mem) (c: t a) (l: list a): Ghost (list (B.pointer (cell a)))\n (requires well_formed h c l)\n (ensures fun _ -> True)\n (decreases l)\n=\n if B.g_is_null c then\n []\n else\n c :: cells h (B.deref h c).next (List.Tot.tl l)\n\n/// A missing bit of automated reasoning\n/// ------------------------------------\n\n/// This should be absolutely trivial, yet, because there's no pattern on\n/// null_unique, we have to do the case analysis and the decomposition\n/// ourselves. This is quite tedious, as it usually takes a while (at least it\n/// did for me!) to realize that this property does not hold automatically.\n///\n/// I'm setting an SMTPat on this lemma because it's absolutely impossible to\n/// get any work done without it.\nlet same_cells_same_pointer #a (h0 h1: HS.mem) (ll0 ll1: t a) (l0 l1: list a): Lemma\n (requires (\n well_formed h0 ll0 l0 /\\\n well_formed h1 ll1 l1 /\\\n cells h0 ll0 l0 == cells h1 ll1 l1))\n (ensures (\n ll0 == ll1))\n [ SMTPat (cells h0 ll0 l0); SMTPat (cells h1 ll1 l1) ]\n=\n if B.g_is_null ll0 && B.g_is_null ll1 then begin\n B.null_unique ll0;\n B.null_unique ll1\n end else if not (B.g_is_null ll0) && not (B.g_is_null ll1) then begin\n ()\n end else\n false_elim ()\n\n/// Classic stateful reasoning lemmas\n/// ---------------------------------\n///\n/// This essentially follows ``stateful``, the type class of a mutable piece of\n/// state defined using the modifies-theory of Low*v2 and equipped with the\n/// right lemmas to enable usage by clients. See\n/// hacl-star/code/streaming/Hacl.Streaming.Interface.fst for more context.\n\nval footprint: (#a: Type) -> (h: HS.mem) -> (l: t a) -> (n: list a) -> Ghost B.loc\n (requires (well_formed h l n))\n (ensures (fun refs -> True))\n (decreases n)\n\nlet rec footprint #a h l n =\n if B.g_is_null l\n then B.loc_none\n else\n let {next = next} = B.get h l 0 in\n let refs = footprint h next (G.hide (L.tl n)) in\n B.loc_union (B.loc_addr_of_buffer l) refs\n\n/// Departing from LinkedList4 here. I prefer to bolt these into the invariant\n/// rather than requiring on some wizard lemmas from LowStar.Monotonic.Buffer to\n/// deduce these.\nval cells_pairwise_disjoint: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_pairwise_disjoint #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.loc_disjoint (B.loc_addr_of_buffer l) (footprint h next (L.tl n)) /\\\n cells_pairwise_disjoint h next (L.tl n)\n\nval cells_live_freeable: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_live_freeable #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.live h l /\\ B.freeable l /\\\n cells_live_freeable h next (L.tl n)\n\nval invariant: #a:Type -> h:HS.mem -> l: t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\nlet invariant #a h l n =\n cells_live_freeable h l n /\\ cells_pairwise_disjoint h l n\n\n/// Normally this would be automatic and writing a custom lemma for that stuff\n/// would be needed only when writing an fsti, however, as stated earlier,\n/// recursion makes everything more difficult and this absolutely needs to be in\n/// scope otherwise the modifies-clause theory cannot deduce disjointness of\n/// fresh allocations w.r.t. the footprint.\nlet rec invariant_loc_in_footprint (#a: Type) (h: HS.mem) (l: t a) (n: list a): Lemma\n (requires well_formed h l n)\n (ensures B.loc_in (footprint h l n) h)\n (decreases n)\n [ SMTPat (invariant h l n) ]\n=\n if B.g_is_null l then\n ()\n else\n let next = (B.deref h l).next in\n invariant_loc_in_footprint h next (L.tl n)\n\n/// Another absolutely essential lemma, for which interestingly the pattern\n/// \"invariant h1 l n\" does not work, but well_formed does... I guess that's\n/// alright.\nlet rec frame (#a: Type) (l: t a) (n: list a) (r: B.loc) (h0 h1: HS.mem): Lemma\n (requires (\n well_formed h0 l n /\\\n invariant h0 l n /\\\n B.loc_disjoint r (footprint h0 l n) /\\\n B.modifies r h0 h1\n ))\n (ensures (\n well_formed h1 l n /\\\n footprint h1 l n == footprint h0 l n /\\\n cells h1 l n == cells h0 l n /\\\n invariant h1 l n\n ))\n (decreases n)\n [ SMTPat (well_formed h1 l n); SMTPat (B.modifies r h0 h1) ]\n=\n if B.g_is_null l then\n ()\n else\n frame (B.deref h0 l).next (L.tl n) r h0 h1\n\n/// Lemmas for working with linked lists\n/// ------------------------------------\n///\n/// Since it's a recursive data structure, a lot of the automated reasoning from\n/// LowStar.Buffer doesn't work since it requires inductions. Some helpful\n/// properties are therefore proved here.\n\nlet rec length_functional #a (h: HS.mem) (c: t a) (l1 l2: list a):\n Lemma\n (requires (well_formed h c l1 /\\ well_formed h c l2))\n (ensures (l1 == l2))\n (decreases (G.reveal l1))\n [ SMTPat (well_formed h c l1); SMTPat (well_formed h c l2) ] =\n if B.g_is_null c\n then ()\n else\n let { next=next } = B.get h c 0 in\n length_functional h next (G.hide (L.tl l1)) (G.hide (L.tl l2))\n\n/// The footprint is based on loc_addr_of_buffer so that we can write a free\n/// operation that operates on the footprint. However, this invalidates most of\n/// the helper lemmas for disjointness that were present in the previous\n/// iteration of this module, named LinkedList4.fst. It's ok, we bake\n/// disjointness directly onto the invariant and it works just as well.\n\n/// Connection between ``cells`` and ``v``\n/// --------------------------------------\n\n/// Redefining some helpers.\n#push-options \"--ifuel 1\"\nlet rec gmap #a #b (f: a -> GTot b) (xs: list a): GTot (list b) =\n match xs with\n | [] -> []\n | x :: xs -> f x :: gmap f xs\n\n/// \"As we all know\", right folds are the easiest to work with because they\n/// follow the natural structure of recursion (left-folds are evil).\nlet rec gfold_right #a #b (f: b -> a -> GTot b) (xs: list a) (acc: b): Ghost b\n (requires True)\n (ensures fun _ -> True)\n (decreases xs)\n=\n match xs with\n | [] -> acc\n | x :: xs -> f (gfold_right f xs acc) x\n#pop-options\n\n/// Connecting ``v`` and ``cells``.\n///\n/// See ``tests/Wireguard.fst`` for why going through gmap is important and how\n/// to structure your definitions (should you need to reason about ``cells``) in\n/// a way that will make your life easier.\nlet deref_data #a (h: HS.mem) (c: B.pointer (cell a)): GTot a =\n (B.deref h c).data\n\n#push-options \"--fuel 1 --ifuel 1\"\nlet rec deref_cells_is_v #a (h: HS.mem) (ll: t a) (l: list a): Lemma\n (requires\n well_formed h ll l /\\\n invariant h ll l)\n (ensures\n gmap (deref_data h) (cells h ll l) == l)\n (decreases l)\n [ SMTPat (well_formed h ll l) ]\n=\n if B.g_is_null ll then\n ()\n else\n deref_cells_is_v h (B.deref h ll).next (List.Tot.tl l)\n#pop-options\n\n/// Connecting ``footprint`` and ``cells``.\n///\n/// This allows conveniently switching to a low-level representation of the\n/// footprint in case clients want to do some ultra-precise reasoning about\n/// what's happening with the list spine. However, the fact that all\n/// operations from this module specify things in terms of ``cells`` along with\n/// ``same_cells_same_pointer`` should cover most common cases. (This predates\n/// the introduction of ``same_cells_same_pointer``.)\n///\n/// So, no SMTPat here. Note, I also used the same trick as above with\n/// higher-order combinators to isolate the use of the argument ``h`` to a\n/// single sub-term rather than having it passed down the recursive calls.\n#push-options \"--fuel 2 --ifuel 2\"\nlet rec footprint_via_cells_is_footprint #a (h: HS.mem) (ll: t a) (l: list a): Lemma\n (requires well_formed h ll l)\n (ensures\n footprint h ll l == (gfold_right B.loc_union (gmap B.loc_addr_of_buffer (cells h ll l)) B.loc_none))\n (decreases l)\n=\n if B.g_is_null ll then\n ()\n else\n footprint_via_cells_is_footprint h (B.deref h ll).next (List.Tot.tl l)\n#pop-options\n\n\n/// Stateful operations\n/// -------------------\n///\n/// One thing to note: we pass the region `r` explicitly, which is certainly not\n/// very convenient, but that's alright, another layer of abstraction will take\n/// care of adding existentials and extra machinery to make this more pleasant\n/// to use.\n///\n/// All of the operations below are low-level (in the sense that they rely on\n/// the predicate). I expect clients to use exclusively the variants of these\n/// functions present in LL2.\n\nval push: (#a: Type) -> (r: HS.rid) -> (n: G.erased (list a)) -> (pl: B.pointer (t a)) -> (x: a) ->\n ST unit\n (requires (fun h ->\n let l = B.deref h pl in\n B.live h pl /\\\n well_formed h l n /\\\n invariant h l n /\\\n ST.is_eternal_region r /\\\n B.(loc_includes (loc_region_only true r) (footprint h l n)) /\\\n B.(loc_disjoint (loc_buffer pl) (loc_region_only true r))\n ))\n (ensures (fun h0 _ h1 ->\n let n' = G.hide (x :: G.reveal n) in\n let l = B.deref h1 pl in\n // Style note: I don't repeat ``B.live pl`` in the post-condition since\n // ``B.modifies (loc_buffer pl) h0 h1`` implies that ``B.live h1 pl``.\n B.modifies (B.loc_buffer pl) h0 h1 /\\\n well_formed h1 l n' /\\\n invariant h1 l n' /\\\n B.(loc_includes (loc_region_only true r) (footprint h1 l n') /\\\n Cons? (cells h1 l n') /\\ List.Tot.tail (cells h1 l n') == cells h0 (B.deref h0 pl) n /\\\n B.fresh_loc (B.loc_addr_of_buffer (List.Tot.hd (cells h1 l n'))) h0 h1)\n ))\n\nlet push #a r n pl x =\n (**) let h0 = ST.get () in\n let l = !* pl in\n let c = { data = x; next = l } in\n\n let pc: B.pointer (cell a) = B.malloc r c 1ul in\n (**) let h1 = ST.get () in\n (**) B.(modifies_only_not_unused_in loc_none h0 h1);\n (**) assert B.(loc_disjoint (loc_buffer pc) (footprint h0 l n));\n\n pl *= pc;\n (**) let h2 = ST.get () in\n (**) let n' = G.hide (x :: G.reveal n) in\n (**) B.(modifies_trans loc_none h0 h1 (loc_buffer pl) h2);\n (**) assert (well_formed h2 (B.deref h2 pl) n');\n (**) assert (invariant h2 (B.deref h2 pl) n');\n (**) assert ((B.deref h2 (B.deref h2 pl)).next == l);\n\n ()\n\nval pop: (#a: Type) -> (r: HS.rid) -> (n: G.erased (list a)) -> (pl: B.pointer (t a)) ->\n ST a\n (requires (fun h ->\n let l = B.deref h pl in\n B.live h pl /\\\n well_formed h l n /\\\n invariant h l n /\\\n ST.is_eternal_region r /\\\n B.(loc_includes (loc_region_only true r) (footprint h l n)) /\\\n B.loc_disjoint (B.loc_buffer pl) (footprint h l n) /\\\n Cons? n\n ))\n (ensures (fun h0 x h1 ->\n let l = B.deref h1 pl in\n let n' = L.tl n in\n x == L.hd n /\\\n // Introducing a super precise modifies clause (e.g. loc_addr_of_buffer\n // (B.deref h0 pl)) here is not useful and prevents trigger-based\n // reasoning, while also requiring clients to unroll footprint. There's a\n // tension here between revealing that ``pop`` does nothing stupid (e.g.\n // re-allocating all the list cells) and providing an abstract enough\n // predicate to work with.\n B.(modifies (loc_buffer pl `loc_union` footprint h0 (B.deref h0 pl) n) h0 h1) /\\\n well_formed h1 l n' /\\\n invariant h1 l n' /\\\n B.(loc_includes (loc_region_only true r) (footprint h1 l n')) /\\\n List.Tot.tail (cells h0 (B.deref h0 pl) n) == cells h1 l n'))\n\nlet pop #a r n pl =\n let l = !* pl in\n let r = (!*l).data in\n let next = (!*l).next in\n (**) let h1 = ST.get () in\n (**) assert (cells h1 l n == l :: cells h1 next (List.Tot.tl n));\n (**) assert B.(loc_disjoint (loc_buffer pl) (footprint h1 l n));\n (**) assert B.(loc_disjoint (loc_buffer pl) (footprint h1 next (List.Tot.tl n)));\n\n pl *= next;\n B.free l;\n r\n\nval free_: (#a: Type) -> (#n: G.erased (list a)) -> (l: t a) ->\n ST unit\n (requires (fun h ->\n well_formed h l n /\\\n invariant h l n\n ))\n (ensures (fun h0 _ h1 ->\n B.(modifies (footprint h0 l n) h0 h1)))\n\nPartial implementation:\nlet rec free_ #a #n l =\n \n"
},
{
"role": "assistant",
"content": "if B.is_null l\nthen ()\nelse\n let tl:G.erased (list a) = G.hide (L.tl (G.reveal n)) in\n free_ #_ #tl (!*l).next;\n B.free l"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n alg: Vale.AES.AES_common_s.algorithm ->\n offset: Prims.int ->\n auth_b: Vale.X64.Memory.buffer128 ->\n abytes_b: Vale.X64.Memory.buffer128 ->\n in128x6_b: Vale.X64.Memory.buffer128 ->\n out128x6_b: Vale.X64.Memory.buffer128 ->\n in128_b: Vale.X64.Memory.buffer128 ->\n out128_b: Vale.X64.Memory.buffer128 ->\n inout_b: Vale.X64.Memory.buffer128 ->\n iv_b: Vale.X64.Memory.buffer128 ->\n iv: Vale.AES.GCM_s.supported_iv_LE ->\n scratch_b: Vale.X64.Memory.buffer128 ->\n key: FStar.Seq.Base.seq Vale.X64.Memory.nat32 ->\n round_keys: FStar.Seq.Base.seq Vale.X64.Decls.quad32 ->\n keys_b: Vale.X64.Memory.buffer128 ->\n hkeys_b: Vale.X64.Memory.buffer128 ->\n expected_tag: FStar.Seq.Base.seq Vale.X64.Memory.nat8 ->\n va_s0: Vale.X64.Decls.va_state ->\n va_k: (_: Vale.X64.Decls.va_state -> _: Prims.unit -> Type0)\n -> Prims.Ghost ((Vale.X64.Decls.va_state * Vale.X64.Decls.va_fuel) * Prims.unit)\nEffect: Prims.Ghost\n\nContext:\nmodule Vale.AES.X64.GCMdecryptOpt\nopen Vale.Def.Prop_s\nopen Vale.Def.Opaque_s\nopen FStar.Seq\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Seq_s\nopen Vale.Def.Types_s\nopen Vale.Arch.Types\nopen Vale.Arch.HeapImpl\nopen Vale.AES.AES_s\nopen Vale.AES.GCTR_s\nopen Vale.AES.GCTR\nopen Vale.AES.GCM\nopen Vale.AES.GHash_s\nopen Vale.AES.GHash\nopen Vale.AES.GCM_s\nopen Vale.AES.X64.AES\nopen Vale.AES.GF128_s\nopen Vale.AES.GF128\nopen Vale.Poly1305.Math\nopen Vale.AES.GCM_helpers\nopen Vale.AES.X64.GHash\nopen Vale.AES.X64.GCTR\nopen Vale.X64.Machine_s\nopen Vale.X64.Memory\nopen Vale.X64.Stack_i\nopen Vale.X64.State\nopen Vale.X64.Decls\nopen Vale.X64.InsBasic\nopen Vale.X64.InsMem\nopen Vale.X64.InsVector\nopen Vale.X64.InsStack\nopen Vale.X64.InsAes\nopen Vale.X64.QuickCode\nopen Vale.X64.QuickCodes\nopen Vale.AES.X64.GF128_Mul\nopen Vale.X64.Stack\nopen Vale.X64.CPU_Features_s\nopen Vale.Math.Poly2.Bits_s\nopen Vale.AES.X64.AESopt\nopen Vale.AES.X64.AESGCM\nopen Vale.AES.X64.AESopt2\nopen Vale.Lib.Meta\nopen Vale.AES.X64.GCMencryptOpt\nopen Vale.AES.OptPublic\nopen Vale.Lib.Basic\n#reset-options \"--z3rlimit 20 --max_ifuel 0\"\n//-- Gcm_extra_bytes\n\nval va_code_Gcm_extra_bytes : alg:algorithm -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_extra_bytes alg =\n (va_Block (va_CCons (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0)\n (va_op_reg_opr64_reg64 rRax) 0 Secret) (va_CCons (va_code_Mov128 (va_op_xmm_xmm 10)\n (va_op_xmm_xmm 0)) (va_CCons (va_code_Ghash_extra_bytes ()) (va_CCons (va_code_Mov128\n (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_CCons (va_code_Pshufb (va_op_xmm_xmm 0)\n (va_op_xmm_xmm 9)) (va_CCons (va_code_AESEncryptBlock alg) (va_CCons (va_code_Pxor\n (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (va_CCons (va_code_Store128_buffer\n (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax) (va_op_xmm_xmm 10) 0 Secret)\n (va_CNil ()))))))))))\n\nval va_codegen_success_Gcm_extra_bytes : alg:algorithm -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_extra_bytes alg =\n (va_pbool_and (va_codegen_success_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0)\n (va_op_reg_opr64_reg64 rRax) 0 Secret) (va_pbool_and (va_codegen_success_Mov128 (va_op_xmm_xmm\n 10) (va_op_xmm_xmm 0)) (va_pbool_and (va_codegen_success_Ghash_extra_bytes ()) (va_pbool_and\n (va_codegen_success_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_pbool_and\n (va_codegen_success_Pshufb (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (va_pbool_and\n (va_codegen_success_AESEncryptBlock alg) (va_pbool_and (va_codegen_success_Pxor (va_op_xmm_xmm\n 10) (va_op_xmm_xmm 0)) (va_pbool_and (va_codegen_success_Store128_buffer\n (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax) (va_op_xmm_xmm 10) 0 Secret)\n (va_ttrue ())))))))))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_extra_bytes (va_mods:va_mods_t) (alg:algorithm) (inout_b:buffer128) (key:(seq\n nat32)) (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (total_bytes:nat)\n (old_hash:quad32) (completed_quads:(seq quad32)) (h_LE:quad32) : (va_quickCode unit\n (va_code_Gcm_extra_bytes alg)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in let (len:(va_int_range\n 1 1)) = 1 in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 188 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64\n rRax) 0 Secret inout_b 0) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 189 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov128 (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (fun (va_s:va_state) _ -> let\n (hash_input:quad32) = va_get_xmm 0 va_s in va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 193 column 22 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Ghash_extra_bytes hkeys_b total_bytes old_hash h_LE completed_quads) (fun\n (va_s:va_state) _ -> va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 194 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (FStar.Seq.Base.equal #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s)\n inout_b) (FStar.Seq.Base.create #quad32 1 hash_input)) (let (snap:(FStar.Seq.Base.seq\n Vale.X64.Decls.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s) inout_b in va_QSeq\n va_range1\n \"***** PRECONDITION NOT MET AT line 198 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 199 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Pshufb (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (fun (va_s:va_state) _ -> va_QBind\n va_range1\n \"***** PRECONDITION NOT MET AT line 200 column 20 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_AESEncryptBlock alg (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 11 va_s)) key\n round_keys keys_b) (fun (va_s:va_state) _ -> va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 201 column 26 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.AES_s.aes_encrypt_LE_reveal ()) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 204 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Pxor (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 205 column 20 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Store128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax)\n (va_op_xmm_xmm 10) 0 Secret inout_b 0) (fun (va_s:va_state) _ -> va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 207 column 24 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.gctr_partial_reveal ()) (va_QEmpty (()))))))))))))))\n\n\nval va_lemma_Gcm_extra_bytes : va_b0:va_code -> va_s0:va_state -> alg:algorithm ->\n inout_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128 -> total_bytes:nat -> old_hash:quad32 -> completed_quads:(seq quad32) ->\n h_LE:quad32\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_extra_bytes alg) va_s0 /\\ va_get_ok va_s0 /\\ (let\n (len:(va_int_range 1 1)) = 1 in sse_enabled /\\ Vale.X64.Decls.buffers_disjoint128 keys_b\n inout_b /\\ Vale.X64.Decls.buffers_disjoint128 hkeys_b inout_b /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) (va_get_reg64 rRax va_s0) inout_b\n len (va_get_mem_layout va_s0) Secret /\\ len == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 inout_b /\\ va_get_xmm 9 va_s0 == Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\ aes_reqs alg key round_keys\n keys_b (va_get_reg64 rR8 va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\\n pclmulqdq_enabled /\\ Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0\n va_s0) hkeys_b) (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret /\\ va_get_xmm 8 va_s0 == Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.AES.GHash.ghash_incremental0 h_LE old_hash completed_quads) /\\ FStar.Seq.Base.length\n #quad32 completed_quads == total_bytes `op_Division` 16 /\\ total_bytes < 16 `op_Multiply`\n FStar.Seq.Base.length #quad32 completed_quads + 16 /\\ va_get_reg64 rR10 va_s0 == total_bytes\n `op_Modulus` 16 /\\ total_bytes `op_Modulus` 16 =!= 0 /\\ (0 < total_bytes /\\ total_bytes < 16\n `op_Multiply` Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes) /\\ 16 `op_Multiply`\n (Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes - 1) < total_bytes)))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (let (len:(va_int_range 1 1)) = 1 in Vale.X64.Decls.modifies_buffer128 inout_b\n (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5 va_sM) /\\ Vale.AES.GCTR.gctr_partial alg len\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b) key (va_get_xmm 11 va_s0) /\\ (let raw_quads =\n FStar.Seq.Base.append #quad32 completed_quads (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0)\n inout_b) in let input_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads) 0 total_bytes in let padded_bytes =\n Vale.AES.GCTR_s.pad_to_128_bits input_bytes in let input_quads =\n Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in l_and (FStar.Seq.Base.length\n #Vale.Def.Types_s.quad32 input_quads > 0) (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8\n va_sM) == Vale.AES.GHash.ghash_incremental h_LE old_hash input_quads))) /\\ va_state_eq va_sM\n (va_update_flags va_sM (va_update_mem_heaplet 5 va_sM (va_update_xmm 10 va_sM (va_update_xmm 8\n va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5 va_sM (va_update_xmm 4\n va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1 va_sM (va_update_xmm 0\n va_sM (va_update_reg64 rR11 va_sM (va_update_reg64 rRcx va_sM (va_update_ok va_sM\n (va_update_mem va_sM va_s0))))))))))))))))))\n[@\"opaque_to_smt\"]\nlet va_lemma_Gcm_extra_bytes va_b0 va_s0 alg inout_b key round_keys keys_b hkeys_b total_bytes\n old_hash completed_quads h_LE =\n let (va_mods:va_mods_t) = [va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8;\n va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm\n 1; va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_ok; va_Mod_mem] in\n let va_qc = va_qcode_Gcm_extra_bytes va_mods alg inout_b key round_keys keys_b hkeys_b\n total_bytes old_hash completed_quads h_LE in\n let (va_sM, va_fM, va_g) = va_wp_sound_code_norm (va_code_Gcm_extra_bytes alg) va_qc va_s0 (fun\n va_s0 va_sM va_g -> let () = va_g in label va_range1\n \"***** POSTCONDITION NOT MET AT line 121 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\ (let (len:(va_int_range 1 1)) = 1 in label va_range1\n \"***** POSTCONDITION NOT MET AT line 174 column 55 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 177 column 95 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.AES.GCTR.gctr_partial alg len (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM) inout_b) key (va_get_xmm 11 va_s0)) /\\ label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 180 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let raw_quads = FStar.Seq.Base.append #quad32 completed_quads (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 181 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let input_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads) 0 total_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 182 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let padded_bytes = Vale.AES.GCTR_s.pad_to_128_bits input_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 183 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let input_quads = Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 186 column 59 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (l_and (FStar.Seq.Base.length #Vale.Def.Types_s.quad32 input_quads > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE old_hash input_quads)))))))) in\n assert_norm (va_qc.mods == va_mods);\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm\n 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1;\n va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_ok; va_Mod_mem]) va_sM va_s0;\n (va_sM, va_fM)\n\n[@ va_qattr]\nlet va_wp_Gcm_extra_bytes (alg:algorithm) (inout_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (total_bytes:nat) (old_hash:quad32)\n (completed_quads:(seq quad32)) (h_LE:quad32) (va_s0:va_state) (va_k:(va_state -> unit -> Type0))\n : Type0 =\n (va_get_ok va_s0 /\\ (let (len:(va_int_range 1 1)) = 1 in sse_enabled /\\\n Vale.X64.Decls.buffers_disjoint128 keys_b inout_b /\\ Vale.X64.Decls.buffers_disjoint128 hkeys_b\n inout_b /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) (va_get_reg64 rRax\n va_s0) inout_b len (va_get_mem_layout va_s0) Secret /\\ len == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 inout_b /\\ va_get_xmm 9 va_s0 == Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\ aes_reqs alg key round_keys\n keys_b (va_get_reg64 rR8 va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\\n pclmulqdq_enabled /\\ Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0\n va_s0) hkeys_b) (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret /\\ va_get_xmm 8 va_s0 == Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.AES.GHash.ghash_incremental0 h_LE old_hash completed_quads) /\\ FStar.Seq.Base.length\n #quad32 completed_quads == total_bytes `op_Division` 16 /\\ total_bytes < 16 `op_Multiply`\n FStar.Seq.Base.length #quad32 completed_quads + 16 /\\ va_get_reg64 rR10 va_s0 == total_bytes\n `op_Modulus` 16 /\\ total_bytes `op_Modulus` 16 =!= 0 /\\ (0 < total_bytes /\\ total_bytes < 16\n `op_Multiply` Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes) /\\ 16 `op_Multiply`\n (Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes - 1) < total_bytes) /\\ (forall\n (va_x_mem:vale_heap) (va_x_rcx:nat64) (va_x_r11:nat64) (va_x_xmm0:quad32) (va_x_xmm1:quad32)\n (va_x_xmm2:quad32) (va_x_xmm3:quad32) (va_x_xmm4:quad32) (va_x_xmm5:quad32) (va_x_xmm6:quad32)\n (va_x_xmm7:quad32) (va_x_xmm8:quad32) (va_x_xmm10:quad32) (va_x_heap5:vale_heap)\n (va_x_efl:Vale.X64.Flags.t) . let va_sM = va_upd_flags va_x_efl (va_upd_mem_heaplet 5\n va_x_heap5 (va_upd_xmm 10 va_x_xmm10 (va_upd_xmm 8 va_x_xmm8 (va_upd_xmm 7 va_x_xmm7\n (va_upd_xmm 6 va_x_xmm6 (va_upd_xmm 5 va_x_xmm5 (va_upd_xmm 4 va_x_xmm4 (va_upd_xmm 3 va_x_xmm3\n (va_upd_xmm 2 va_x_xmm2 (va_upd_xmm 1 va_x_xmm1 (va_upd_xmm 0 va_x_xmm0 (va_upd_reg64 rR11\n va_x_r11 (va_upd_reg64 rRcx va_x_rcx (va_upd_mem va_x_mem va_s0)))))))))))))) in va_get_ok\n va_sM /\\ (let (len:(va_int_range 1 1)) = 1 in Vale.X64.Decls.modifies_buffer128 inout_b\n (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5 va_sM) /\\ Vale.AES.GCTR.gctr_partial alg len\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b) key (va_get_xmm 11 va_s0) /\\ (let raw_quads =\n FStar.Seq.Base.append #quad32 completed_quads (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0)\n inout_b) in let input_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads) 0 total_bytes in let padded_bytes =\n Vale.AES.GCTR_s.pad_to_128_bits input_bytes in let input_quads =\n Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in l_and (FStar.Seq.Base.length\n #Vale.Def.Types_s.quad32 input_quads > 0) (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8\n va_sM) == Vale.AES.GHash.ghash_incremental h_LE old_hash input_quads))) ==> va_k va_sM (())))\n\n\nval va_wpProof_Gcm_extra_bytes : alg:algorithm -> inout_b:buffer128 -> key:(seq nat32) ->\n round_keys:(seq quad32) -> keys_b:buffer128 -> hkeys_b:buffer128 -> total_bytes:nat ->\n old_hash:quad32 -> completed_quads:(seq quad32) -> h_LE:quad32 -> va_s0:va_state ->\n va_k:(va_state -> unit -> Type0)\n -> Ghost (va_state & va_fuel & unit)\n (requires (va_t_require va_s0 /\\ va_wp_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b\n total_bytes old_hash completed_quads h_LE va_s0 va_k))\n (ensures (fun (va_sM, va_f0, va_g) -> va_t_ensure (va_code_Gcm_extra_bytes alg) ([va_Mod_flags;\n va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5;\n va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR11;\n va_Mod_reg64 rRcx; va_Mod_mem]) va_s0 va_k ((va_sM, va_f0, va_g))))\n[@\"opaque_to_smt\"]\nlet va_wpProof_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b total_bytes old_hash\n completed_quads h_LE va_s0 va_k =\n let (va_sM, va_f0) = va_lemma_Gcm_extra_bytes (va_code_Gcm_extra_bytes alg) va_s0 alg inout_b key\n round_keys keys_b hkeys_b total_bytes old_hash completed_quads h_LE in\n va_lemma_upd_update va_sM;\n assert (va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 5 va_sM (va_update_xmm 10\n va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5\n va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1\n va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR11 va_sM (va_update_reg64 rRcx va_sM\n (va_update_ok va_sM (va_update_mem va_sM va_s0)))))))))))))))));\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm\n 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1;\n va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_mem]) va_sM va_s0;\n let va_g = () in\n (va_sM, va_f0, va_g)\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_quick_Gcm_extra_bytes (alg:algorithm) (inout_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (total_bytes:nat) (old_hash:quad32)\n (completed_quads:(seq quad32)) (h_LE:quad32) : (va_quickCode unit (va_code_Gcm_extra_bytes alg)) =\n (va_QProc (va_code_Gcm_extra_bytes alg) ([va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10;\n va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm\n 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_mem])\n (va_wp_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b total_bytes old_hash\n completed_quads h_LE) (va_wpProof_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b\n total_bytes old_hash completed_quads h_LE))\n//--\n//-- Gcm_blocks128\n\nval va_code_Gcm_blocks128 : alg:algorithm -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_blocks128 alg =\n (va_Block (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRbx) (va_op_opr64_reg64 rRdi))\n (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR12) (va_op_opr64_reg64 rRdx)) (va_CCons\n (va_code_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRax)) (va_CCons\n (va_code_Ghash_buffer ()) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRdi)\n (va_op_opr64_reg64 rRbx)) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRdx)\n (va_op_opr64_reg64 rR12)) (va_CCons (va_code_Gctr_blocks128 alg) (va_CNil ())))))))))\n\nval va_codegen_success_Gcm_blocks128 : alg:algorithm -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_blocks128 alg =\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rRbx) (va_op_opr64_reg64 rRdi))\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR12) (va_op_opr64_reg64 rRdx))\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRax))\n (va_pbool_and (va_codegen_success_Ghash_buffer ()) (va_pbool_and (va_codegen_success_Mov64\n (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRbx)) (va_pbool_and (va_codegen_success_Mov64\n (va_op_dst_opr64_reg64 rRdx) (va_op_opr64_reg64 rR12)) (va_pbool_and\n (va_codegen_success_Gctr_blocks128 alg) (va_ttrue ()))))))))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_blocks128 (va_mods:va_mods_t) (alg:algorithm) (in_b:buffer128) (out_b:buffer128)\n (key:(seq nat32)) (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (h_LE:quad32)\n : (va_quickCode unit (va_code_Gcm_blocks128 alg)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 274 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRbx) (va_op_opr64_reg64 rRdi)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 275 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR12) (va_op_opr64_reg64 rRdx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 276 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRax)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 277 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Ghash_buffer hkeys_b in_b h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8\n va_old_s))) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 278 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRbx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 279 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRdx) (va_op_opr64_reg64 rR12)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 280 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gctr_blocks128 alg in_b out_b key round_keys keys_b) (va_QEmpty (()))))))))))\n\n\nval va_lemma_Gcm_blocks128 : va_b0:va_code -> va_s0:va_state -> alg:algorithm -> in_b:buffer128 ->\n out_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128 -> h_LE:quad32\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_blocks128 alg) va_s0 /\\ va_get_ok va_s0 /\\\n (sse_enabled /\\ Vale.X64.Decls.buffers_disjoint128 keys_b out_b /\\\n Vale.X64.Decls.buffers_disjoint128 hkeys_b out_b /\\ (Vale.X64.Decls.buffers_disjoint128 in_b\n out_b \\/ in_b == out_b) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0)\n (va_get_reg64 rRax va_s0) in_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1 va_s0) (va_get_reg64 rRdi va_s0) out_b\n (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\ va_get_reg64 rRax va_s0 + 16\n `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply`\n va_get_reg64 rRdx va_s0 < pow2_64 /\\ l_and (Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 in_b == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out_b)\n (Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b < pow2_32) /\\ va_get_reg64 rRdx\n va_s0 == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b /\\ va_get_xmm 9 va_s0 ==\n Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\\n va_get_reg64 rRdx va_s0 < pow2_32 /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rR8\n va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret)))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (Vale.X64.Decls.modifies_buffer128 out_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM) /\\ Vale.AES.GCTR.gctr_partial alg (va_get_reg64 rRdx va_s0) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b) key\n (va_get_xmm 11 va_s0) /\\ va_get_xmm 11 va_sM == Vale.AES.GCTR.inc32lite (va_get_xmm 11 va_s0)\n (va_get_reg64 rRdx va_s0) /\\ (va_get_reg64 rRdx va_s0 == 0 ==> l_and (va_get_xmm 8 va_sM ==\n va_get_xmm 8 va_s0) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b ==\n Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) out_b)) /\\ (va_get_reg64 rRdx va_s0 > 0 ==>\n l_and (va_get_reg64 rRdx va_s0 <= FStar.Seq.Base.length #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) ==> FStar.Seq.Base.length\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.slice #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) 0 (va_get_reg64 rRdx va_s0)) > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s0)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b)))) /\\ va_state_eq va_sM (va_update_flags va_sM\n (va_update_mem_heaplet 1 va_sM (va_update_xmm 10 va_sM (va_update_xmm 11 va_sM (va_update_xmm 8\n va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5 va_sM (va_update_xmm 4\n va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1 va_sM (va_update_xmm 0\n va_sM (va_update_reg64 rR12 va_sM (va_update_reg64 rRdx va_sM (va_update_reg64 rR10 va_sM\n (va_update_reg64 rR11 va_sM (va_update_reg64 rRdi va_sM (va_update_reg64 rRbx va_sM\n (va_update_ok va_sM (va_update_mem va_sM va_s0)))))))))))))))))))))))\n[@\"opaque_to_smt\"]\nlet va_lemma_Gcm_blocks128 va_b0 va_s0 alg in_b out_b key round_keys keys_b hkeys_b h_LE =\n let (va_mods:va_mods_t) = [va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11;\n va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm\n 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10;\n va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_ok; va_Mod_mem] in\n let va_qc = va_qcode_Gcm_blocks128 va_mods alg in_b out_b key round_keys keys_b hkeys_b h_LE in\n let (va_sM, va_fM, va_g) = va_wp_sound_code_norm (va_code_Gcm_blocks128 alg) va_qc va_s0 (fun\n va_s0 va_sM va_g -> let () = va_g in label va_range1\n \"***** POSTCONDITION NOT MET AT line 210 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\ (label va_range1\n \"***** POSTCONDITION NOT MET AT line 255 column 53 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 261 column 95 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.AES.GCTR.gctr_partial alg (va_get_reg64 rRdx va_s0) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b) key\n (va_get_xmm 11 va_s0)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 262 column 45 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_xmm 11 va_sM == Vale.AES.GCTR.inc32lite (va_get_xmm 11 va_s0) (va_get_reg64 rRdx\n va_s0)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 265 column 93 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRdx va_s0 == 0 ==> l_and (va_get_xmm 8 va_sM == va_get_xmm 8 va_s0)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b == Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) out_b)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 267 column 131 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRdx va_s0 > 0 ==> l_and (va_get_reg64 rRdx va_s0 <= FStar.Seq.Base.length\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) ==>\n FStar.Seq.Base.length #Vale.X64.Decls.quad32 (FStar.Seq.Base.slice #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) 0 (va_get_reg64 rRdx va_s0)) > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s0)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b))))) in\n assert_norm (va_qc.mods == va_mods);\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11; va_Mod_xmm\n 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2;\n va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10;\n va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_ok; va_Mod_mem]) va_sM va_s0;\n (va_sM, va_fM)\n\n[@ va_qattr]\nlet va_wp_Gcm_blocks128 (alg:algorithm) (in_b:buffer128) (out_b:buffer128) (key:(seq nat32))\n (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (h_LE:quad32) (va_s0:va_state)\n (va_k:(va_state -> unit -> Type0)) : Type0 =\n (va_get_ok va_s0 /\\ (sse_enabled /\\ Vale.X64.Decls.buffers_disjoint128 keys_b out_b /\\\n Vale.X64.Decls.buffers_disjoint128 hkeys_b out_b /\\ (Vale.X64.Decls.buffers_disjoint128 in_b\n out_b \\/ in_b == out_b) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0)\n (va_get_reg64 rRax va_s0) in_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1 va_s0) (va_get_reg64 rRdi va_s0) out_b\n (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\ va_get_reg64 rRax va_s0 + 16\n `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply`\n va_get_reg64 rRdx va_s0 < pow2_64 /\\ l_and (Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 in_b == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out_b)\n (Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b < pow2_32) /\\ va_get_reg64 rRdx\n va_s0 == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b /\\ va_get_xmm 9 va_s0 ==\n Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\\n va_get_reg64 rRdx va_s0 < pow2_32 /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rR8\n va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret) /\\ (forall (va_x_mem:vale_heap) (va_x_rbx:nat64) (va_x_rdi:nat64) (va_x_r11:nat64)\n (va_x_r10:nat64) (va_x_rdx:nat64) (va_x_r12:nat64) (va_x_xmm0:quad32) (va_x_xmm1:quad32)\n (va_x_xmm2:quad32) (va_x_xmm3:quad32) (va_x_xmm4:quad32) (va_x_xmm5:quad32) (va_x_xmm6:quad32)\n (va_x_xmm7:quad32) (va_x_xmm8:quad32) (va_x_xmm11:quad32) (va_x_xmm10:quad32)\n (va_x_heap1:vale_heap) (va_x_efl:Vale.X64.Flags.t) . let va_sM = va_upd_flags va_x_efl\n (va_upd_mem_heaplet 1 va_x_heap1 (va_upd_xmm 10 va_x_xmm10 (va_upd_xmm 11 va_x_xmm11\n (va_upd_xmm 8 va_x_xmm8 (va_upd_xmm 7 va_x_xmm7 (va_upd_xmm 6 va_x_xmm6 (va_upd_xmm 5 va_x_xmm5\n (va_upd_xmm 4 va_x_xmm4 (va_upd_xmm 3 va_x_xmm3 (va_upd_xmm 2 va_x_xmm2 (va_upd_xmm 1 va_x_xmm1\n (va_upd_xmm 0 va_x_xmm0 (va_upd_reg64 rR12 va_x_r12 (va_upd_reg64 rRdx va_x_rdx (va_upd_reg64\n rR10 va_x_r10 (va_upd_reg64 rR11 va_x_r11 (va_upd_reg64 rRdi va_x_rdi (va_upd_reg64 rRbx\n va_x_rbx (va_upd_mem va_x_mem va_s0))))))))))))))))))) in va_get_ok va_sM /\\\n (Vale.X64.Decls.modifies_buffer128 out_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM) /\\ Vale.AES.GCTR.gctr_partial alg (va_get_reg64 rRdx va_s0) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b) key\n (va_get_xmm 11 va_s0) /\\ va_get_xmm 11 va_sM == Vale.AES.GCTR.inc32lite (va_get_xmm 11 va_s0)\n (va_get_reg64 rRdx va_s0) /\\ (va_get_reg64 rRdx va_s0 == 0 ==> l_and (va_get_xmm 8 va_sM ==\n va_get_xmm 8 va_s0) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b ==\n Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) out_b)) /\\ (va_get_reg64 rRdx va_s0 > 0 ==>\n l_and (va_get_reg64 rRdx va_s0 <= FStar.Seq.Base.length #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) ==> FStar.Seq.Base.length\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.slice #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) 0 (va_get_reg64 rRdx va_s0)) > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s0)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b)))) ==> va_k va_sM (())))\n\n\nval va_wpProof_Gcm_blocks128 : alg:algorithm -> in_b:buffer128 -> out_b:buffer128 -> key:(seq\n nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 -> hkeys_b:buffer128 -> h_LE:quad32 ->\n va_s0:va_state -> va_k:(va_state -> unit -> Type0)\n -> Ghost (va_state & va_fuel & unit)\n (requires (va_t_require va_s0 /\\ va_wp_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b\n h_LE va_s0 va_k))\n (ensures (fun (va_sM, va_f0, va_g) -> va_t_ensure (va_code_Gcm_blocks128 alg) ([va_Mod_flags;\n va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6;\n va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0;\n va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10; va_Mod_reg64 rR11; va_Mod_reg64 rRdi;\n va_Mod_reg64 rRbx; va_Mod_mem]) va_s0 va_k ((va_sM, va_f0, va_g))))\n[@\"opaque_to_smt\"]\nlet va_wpProof_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b h_LE va_s0 va_k =\n let (va_sM, va_f0) = va_lemma_Gcm_blocks128 (va_code_Gcm_blocks128 alg) va_s0 alg in_b out_b key\n round_keys keys_b hkeys_b h_LE in\n va_lemma_upd_update va_sM;\n assert (va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 1 va_sM (va_update_xmm 10\n va_sM (va_update_xmm 11 va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6\n va_sM (va_update_xmm 5 va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2\n va_sM (va_update_xmm 1 va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR12 va_sM\n (va_update_reg64 rRdx va_sM (va_update_reg64 rR10 va_sM (va_update_reg64 rR11 va_sM\n (va_update_reg64 rRdi va_sM (va_update_reg64 rRbx va_sM (va_update_ok va_sM (va_update_mem\n va_sM va_s0))))))))))))))))))))));\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11; va_Mod_xmm\n 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2;\n va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10;\n va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_mem]) va_sM va_s0;\n let va_g = () in\n (va_sM, va_f0, va_g)\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_quick_Gcm_blocks128 (alg:algorithm) (in_b:buffer128) (out_b:buffer128) (key:(seq nat32))\n (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (h_LE:quad32) : (va_quickCode\n unit (va_code_Gcm_blocks128 alg)) =\n (va_QProc (va_code_Gcm_blocks128 alg) ([va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10;\n va_Mod_xmm 11; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm\n 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64\n rR10; va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_mem])\n (va_wp_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b h_LE)\n (va_wpProof_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b h_LE))\n//--\n//-- Gcm_blocks\n#push-options \"--z3rlimit 1000\"\n\nval va_code_Gcm_blocks : alg:algorithm -> offset:int -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_blocks alg offset =\n (va_Block (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR13) (va_op_opr64_reg64 rRcx))\n (va_CCons (va_code_AddLea64 (va_op_dst_opr64_reg64 rR9) (va_op_opr64_reg64 rR9) (va_const_opr64\n 32)) (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rRbx) (va_op_reg_opr64_reg64 rRsp)\n (offset + 0)) (va_CCons (va_code_Gcm_blocks_auth ()) (va_CCons (va_code_Load64_stack\n (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 8)) (va_CCons\n (va_code_Load64_stack (va_op_dst_opr64_reg64 rRsi) (va_op_reg_opr64_reg64 rRsp) (offset + 16))\n (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp)\n (offset + 24)) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRcx) (va_op_opr64_reg64 rR13))\n (va_CCons (va_code_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (va_CCons\n (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 2) (va_op_xmm_xmm 1) (va_op_reg_opr64_reg64\n rR8) 0 Public) (va_CCons (va_code_Store128_buffer (va_op_heaplet_mem_heaplet 3)\n (va_op_reg_opr64_reg64 rRbp) (va_op_xmm_xmm 1) 0 Secret) (va_CCons (va_code_Load_one_lsb\n (va_op_xmm_xmm 10)) (va_CCons (va_code_VPaddd (va_op_xmm_xmm 1) (va_op_xmm_xmm 1)\n (va_op_xmm_xmm 10)) (va_CCons (va_code_AES_GCM_decrypt_6mult alg) (va_CCons\n (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 11) (va_op_reg_opr64_reg64\n rRbp) 32 Secret) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR8) (va_op_opr64_reg64 rRcx))\n (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp)\n (offset + 32)) (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rRdi)\n (va_op_reg_opr64_reg64 rRsp) (offset + 40)) (va_CCons (va_code_Load64_stack\n (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp) (offset + 48)) (va_CCons\n (va_code_Mov64 (va_op_dst_opr64_reg64 rR14) (va_op_opr64_reg64 rRdx)) (va_CCons\n (va_code_InitPshufbMask (va_op_xmm_xmm 9) (va_op_reg_opr64_reg64 rR12)) (va_CCons\n (va_code_Pshufb (va_op_xmm_xmm 11) (va_op_xmm_xmm 9)) (va_CCons (va_code_Gcm_blocks128 alg)\n (va_CCons (va_code_Stack_lemma ()) (va_CCons (va_code_Add64 (va_op_dst_opr64_reg64 rR14)\n (va_opr_code_Stack (va_op_reg64_reg64 rRsp) (offset + 24) Public)) (va_CCons (va_code_IMul64\n (va_op_dst_opr64_reg64 rR14) (va_const_opr64 16)) (va_CCons (va_code_Load64_stack\n (va_op_dst_opr64_reg64 rR13) (va_op_reg_opr64_reg64 rRsp) (offset + 64)) (va_CCons (va_IfElse\n (va_cmp_gt (va_op_cmp_reg64 rR13) (va_op_cmp_reg64 rR14)) (va_Block (va_CCons\n (va_code_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp) (offset + 56))\n (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR10) (va_op_opr64_reg64 rR13)) (va_CCons\n (va_code_And64 (va_op_dst_opr64_reg64 rR10) (va_const_opr64 15)) (va_CCons\n (va_code_Gcm_extra_bytes alg) (va_CCons (va_Block (va_CNil ())) (va_CNil ()))))))) (va_Block\n (va_CNil ()))) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR11) (va_op_opr64_reg64 rR15))\n (va_CCons (va_code_Gcm_make_length_quad ()) (va_CCons (va_code_Ghash_register ()) (va_CCons\n (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64\n rRbp) 0 Secret) (va_CCons (va_code_Gctr_register alg) (va_CCons (va_Block (va_CNil ()))\n (va_CNil ()))))))))))))))))))))))))))))))))))))\n\nval va_codegen_success_Gcm_blocks : alg:algorithm -> offset:int -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_blocks alg offset =\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR13) (va_op_opr64_reg64 rRcx))\n (va_pbool_and (va_codegen_success_AddLea64 (va_op_dst_opr64_reg64 rR9) (va_op_opr64_reg64 rR9)\n (va_const_opr64 32)) (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64\n rRbx) (va_op_reg_opr64_reg64 rRsp) (offset + 0)) (va_pbool_and\n (va_codegen_success_Gcm_blocks_auth ()) (va_pbool_and (va_codegen_success_Load64_stack\n (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 8)) (va_pbool_and\n (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRsi) (va_op_reg_opr64_reg64 rRsp)\n (offset + 16)) (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRdx)\n (va_op_reg_opr64_reg64 rRsp) (offset + 24)) (va_pbool_and (va_codegen_success_Mov64\n (va_op_dst_opr64_reg64 rRcx) (va_op_opr64_reg64 rR13)) (va_pbool_and (va_codegen_success_Mov128\n (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (va_pbool_and (va_codegen_success_Load128_buffer\n (va_op_heaplet_mem_heaplet 2) (va_op_xmm_xmm 1) (va_op_reg_opr64_reg64 rR8) 0 Public)\n (va_pbool_and (va_codegen_success_Store128_buffer (va_op_heaplet_mem_heaplet 3)\n (va_op_reg_opr64_reg64 rRbp) (va_op_xmm_xmm 1) 0 Secret) (va_pbool_and\n (va_codegen_success_Load_one_lsb (va_op_xmm_xmm 10)) (va_pbool_and (va_codegen_success_VPaddd\n (va_op_xmm_xmm 1) (va_op_xmm_xmm 1) (va_op_xmm_xmm 10)) (va_pbool_and\n (va_codegen_success_AES_GCM_decrypt_6mult alg) (va_pbool_and (va_codegen_success_Load128_buffer\n (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 11) (va_op_reg_opr64_reg64 rRbp) 32 Secret)\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR8) (va_op_opr64_reg64 rRcx))\n (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRax)\n (va_op_reg_opr64_reg64 rRsp) (offset + 32)) (va_pbool_and (va_codegen_success_Load64_stack\n (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 40)) (va_pbool_and\n (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp)\n (offset + 48)) (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR14)\n (va_op_opr64_reg64 rRdx)) (va_pbool_and (va_codegen_success_InitPshufbMask (va_op_xmm_xmm 9)\n (va_op_reg_opr64_reg64 rR12)) (va_pbool_and (va_codegen_success_Pshufb (va_op_xmm_xmm 11)\n (va_op_xmm_xmm 9)) (va_pbool_and (va_codegen_success_Gcm_blocks128 alg) (va_pbool_and\n (va_codegen_success_Stack_lemma ()) (va_pbool_and (va_codegen_success_Add64\n (va_op_dst_opr64_reg64 rR14) (va_opr_code_Stack (va_op_reg64_reg64 rRsp) (offset + 24) Public))\n (va_pbool_and (va_codegen_success_IMul64 (va_op_dst_opr64_reg64 rR14) (va_const_opr64 16))\n (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rR13)\n (va_op_reg_opr64_reg64 rRsp) (offset + 64)) (va_pbool_and (va_pbool_and\n (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp)\n (offset + 56)) (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR10)\n (va_op_opr64_reg64 rR13)) (va_pbool_and (va_codegen_success_And64 (va_op_dst_opr64_reg64 rR10)\n (va_const_opr64 15)) (va_codegen_success_Gcm_extra_bytes alg)))) (va_pbool_and\n (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR11) (va_op_opr64_reg64 rR15)) (va_pbool_and\n (va_codegen_success_Gcm_make_length_quad ()) (va_pbool_and (va_codegen_success_Ghash_register\n ()) (va_pbool_and (va_codegen_success_Load128_buffer (va_op_heaplet_mem_heaplet 3)\n (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64 rRbp) 0 Secret) (va_pbool_and\n (va_codegen_success_Gctr_register alg) (va_ttrue ()))))))))))))))))))))))))))))))))))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_blocks (va_mods:va_mods_t) (alg:algorithm) (offset:int) (auth_b:buffer128)\n (abytes_b:buffer128) (in128x6_b:buffer128) (out128x6_b:buffer128) (in128_b:buffer128)\n (out128_b:buffer128) (inout_b:buffer128) (iv_b:buffer128) (scratch_b:buffer128) (key:(seq nat32))\n (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) : (va_quickCode unit\n (va_code_Gcm_blocks alg offset)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in let\n (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 0) (va_get_stack va_s) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 8) (va_get_stack va_s) in let\n (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 16) (va_get_stack va_s) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 24) (va_get_stack va_s) in let\n (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 32) (va_get_stack va_s) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 40) (va_get_stack va_s) in let\n (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset\n + 48) (va_get_stack va_s) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 56) (va_get_stack va_s) in let\n (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s\n + offset + 64) (va_get_stack va_s) in let (h_LE:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.X64.Decls.buffer128_read hkeys_b 2\n (va_get_mem_heaplet 0 va_old_s)) in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 463 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR13) (va_op_opr64_reg64 rRcx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 464 column 13 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_AddLea64 (va_op_dst_opr64_reg64 rR9) (va_op_opr64_reg64 rR9) (va_const_opr64 32))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 465 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRbx) (va_op_reg_opr64_reg64 rRsp) (offset + 0))\n (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 466 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_blocks_auth auth_b abytes_b hkeys_b h_LE) (fun (va_s:va_state)\n (auth_quad_seq:(seq quad32)) -> let (y_0:quad32) = Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 0 0 0 0 in let (y_auth_bytes:quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s) in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 473 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 8))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 474 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRsi) (va_op_reg_opr64_reg64 rRsp) (offset + 16))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 475 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp) (offset + 24))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 476 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRcx) (va_op_opr64_reg64 rR13)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 477 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (fun (va_s:va_state) _ -> let\n (iv_BE:Vale.X64.Decls.quad32) = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2\n va_old_s) in let (ctr_BE_1:quad32) = iv_BE in let (ctr_BE_2:quad32) = Vale.AES.GCTR_s.inc32\n iv_BE 1 in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 483 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 2) (va_op_xmm_xmm 1) (va_op_reg_opr64_reg64\n rR8) 0 Public iv_b 0) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 485 column 20 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Store128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_reg_opr64_reg64 rRbp)\n (va_op_xmm_xmm 1) 0 Secret scratch_b 0) (fun (va_s:va_state) _ -> let (j0:quad32) = va_get_xmm\n 1 va_s in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 487 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load_one_lsb (va_op_xmm_xmm 10)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 489 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_VPaddd (va_op_xmm_xmm 1) (va_op_xmm_xmm 1) (va_op_xmm_xmm 10)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 491 column 26 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_AES_GCM_decrypt_6mult alg h_LE iv_b in128x6_b out128x6_b scratch_b key round_keys\n keys_b hkeys_b) (fun (va_s:va_state) _ -> let (y_cipher128x6:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s) in let (auth_in:(seq quad32)) =\n auth_quad_seq in let (va_arg138:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b in let\n (va_arg137:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = auth_in in let\n (va_arg136:Vale.Def.Types_s.quad32) = y_auth_bytes in let (va_arg135:Vale.Def.Types_s.quad32) =\n y_0 in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 494 column 36 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GHash.lemma_ghash_incremental0_append h_LE va_arg135 va_arg136\n y_cipher128x6 va_arg137 va_arg138) (let auth_in = FStar.Seq.Base.append #quad32 auth_in\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b) in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 498 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 11)\n (va_op_reg_opr64_reg64 rRbp) 32 Secret scratch_b 2) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 499 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR8) (va_op_opr64_reg64 rRcx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 500 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp) (offset + 32))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 501 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 40))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 502 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp) (offset + 48))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 503 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR14) (va_op_opr64_reg64 rRdx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 504 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_InitPshufbMask (va_op_xmm_xmm 9) (va_op_reg_opr64_reg64 rR12)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 505 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Pshufb (va_op_xmm_xmm 11) (va_op_xmm_xmm 9)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 506 column 18 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_blocks128 alg in128_b out128_b key round_keys keys_b hkeys_b h_LE) (fun\n (va_s:va_state) _ -> let (y_cipher128:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s) in let (va_arg134:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b in let\n (va_arg133:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = auth_in in let\n (va_arg132:Vale.Def.Types_s.quad32) = y_0 in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 508 column 36 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GHash.lemma_ghash_incremental0_append h_LE va_arg132 y_cipher128x6\n y_cipher128 va_arg133 va_arg134) (let auth_in = FStar.Seq.Base.append #quad32 auth_in\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b) in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 512 column 22 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Stack_lemma (va_op_reg64_reg64 rRsp) (offset + 24) Public) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 512 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Add64 (va_op_dst_opr64_reg64 rR14) (va_opr_code_Stack (va_op_reg64_reg64 rRsp)\n (offset + 24) Public)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 513 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_IMul64 (va_op_dst_opr64_reg64 rR14) (va_const_opr64 16)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 514 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rR13) (va_op_reg_opr64_reg64 rRsp) (offset + 64))\n (fun (va_s:va_state) _ -> let (y_inout:Vale.Def.Types_s.quad32) = y_cipher128 in let\n (plain_byte_seq:(seq quad32)) = empty_seq_quad32 in let (cipher_byte_seq:(seq quad32)) =\n empty_seq_quad32 in let (va_arg131:Vale.Def.Types_s.quad32) = va_get_xmm 11 va_s in let\n (va_arg130:(FStar.Seq.Base.seq Vale.Def.Types_s.nat32)) = key in let\n (va_arg129:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = cipher_byte_seq in let\n (va_arg128:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = plain_byte_seq in let\n (va_arg127:Vale.AES.AES_common_s.algorithm) = alg in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 519 column 29 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.gctr_partial_opaque_init va_arg127 va_arg128 va_arg129 va_arg130\n va_arg131) (let (total_bytes:(va_int_at_least 0)) = FStar.Seq.Base.length #quad32 auth_quad_seq\n `op_Multiply` 16 + plain_num_bytes in va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 523 column 8 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_qIf va_mods (Cmp_gt (va_op_cmp_reg64 rR13) (va_op_cmp_reg64 rR14)) (qblock va_mods (fun\n (va_s:va_state) -> va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 525 column 21 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp) (offset + 56))\n (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 526 column 14 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR10) (va_op_opr64_reg64 rR13)) (fun (va_s:va_state) _\n -> va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 527 column 26 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.Poly1305.Math.lemma_poly_bits64 ()) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 528 column 14 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_And64 (va_op_dst_opr64_reg64 rR10) (va_const_opr64 15)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 532 column 24 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b total_bytes y_0 auth_in\n h_LE) (fun (va_s:va_state) _ -> let y_inout = Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm\n 8 va_s) in let (raw_auth_quads:(FStar.Seq.Base.seq quad32)) = FStar.Seq.Base.append #quad32\n auth_in (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_old_s) inout_b) in va_qAssertSquash\n va_range1\n \"***** EXPRESSION PRECONDITIONS NOT MET WITHIN line 536 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n ((fun a_1906 (s_1907:(FStar.Seq.Base.seq a_1906)) (i_1908:Prims.nat) (j_1909:Prims.nat) -> let\n (j_1869:Prims.nat) = j_1909 in Prims.b2t (Prims.op_AmpAmp (Prims.op_LessThanOrEqual i_1908\n j_1869) (Prims.op_LessThanOrEqual j_1869 (FStar.Seq.Base.length #a_1906 s_1907))))\n Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_auth_quads) 0 total_bytes)\n (fun _ -> let (auth_input_bytes:(FStar.Seq.Base.seq Vale.Def.Types_s.nat8)) =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n raw_auth_quads) 0 total_bytes in let (padded_auth_bytes:(FStar.Seq.Base.seq\n Vale.Def.Types_s.nat8)) = Vale.AES.GCTR_s.pad_to_128_bits auth_input_bytes in let auth_in =\n Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_auth_bytes in let plain_byte_seq =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_old_s) inout_b in let cipher_byte_seq =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s) inout_b in va_QEmpty ((auth_in,\n cipher_byte_seq, plain_byte_seq, y_inout)))))))))) (qblock va_mods (fun (va_s:va_state) ->\n va_QEmpty ((auth_in, cipher_byte_seq, plain_byte_seq, y_inout))))) (fun (va_s:va_state) va_g ->\n let ((auth_in:(seq quad32)), (cipher_byte_seq:(seq quad32)), (plain_byte_seq:(seq quad32)),\n (y_inout:Vale.Def.Types_s.quad32)) = va_g in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 547 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR11) (va_op_opr64_reg64 rR15)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 548 column 25 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_make_length_quad ()) (fun (va_s:va_state) _ -> let\n (length_quad32:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 0\n va_s) in va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 551 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Ghash_register hkeys_b h_LE y_inout) (fun (va_s:va_state) _ -> let\n (y_final:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s)\n in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 554 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64\n rRbp) 0 Secret scratch_b 0) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 557 column 18 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gctr_register alg key round_keys keys_b) (fun (va_s:va_state) _ -> let\n (va_arg126:Vale.Def.Types_s.quad32) = va_get_xmm 8 va_s in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 560 column 40 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.Arch.Types.le_seq_quad32_to_bytes_of_singleton va_arg126)\n (va_qAssertSquash va_range1\n \"***** EXPRESSION PRECONDITIONS NOT MET WITHIN line 561 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n ((fun (icb_BE_677:Vale.Def.Types_s.quad32) (plain_LE_678:Vale.Def.Types_s.quad32)\n (alg_679:Vale.AES.AES_common_s.algorithm) (key_680:(FStar.Seq.Base.seq Vale.Def.Types_s.nat32))\n (i_681:Prims.int) -> Vale.AES.AES_s.is_aes_key_LE alg_679 key_680) j0 y_final alg key 0) (fun _\n -> va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 561 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_xmm 8 va_s == Vale.AES.GCTR_s.gctr_encrypt_block j0 y_final alg key 0) (let\n (plain128:(FStar.Seq.Base.seq Vale.X64.Decls.quad32)) = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b) in let\n (cipher128:(FStar.Seq.Base.seq Vale.X64.Decls.quad32)) = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) in128_b) in va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 566 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (FStar.Seq.Base.length #quad32 plain_byte_seq == 0 ==> FStar.Seq.Base.equal\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 plain128 plain_byte_seq)\n plain128) (va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 567 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (FStar.Seq.Base.length #quad32 cipher_byte_seq == 0 ==> FStar.Seq.Base.equal\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 cipher128 cipher_byte_seq)\n cipher128) (let (va_arg125:Vale.Def.Types_s.quad32) = Vale.AES.GCTR.inc32lite ctr_BE_2 len128x6\n in let (va_arg124:Vale.Def.Types_s.quad32) = ctr_BE_2 in let (va_arg123:(FStar.Seq.Base.seq\n Vale.Def.Types_s.nat32)) = key in let (va_arg122:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32))\n = Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) out128_b in let\n (va_arg121:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_old_s) in128_b in let (va_arg120:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s) out128x6_b in let\n (va_arg119:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_old_s) in128x6_b in let (va_arg118:Prims.nat) = len128 in let\n (va_arg117:Prims.nat) = len128x6 in let (va_arg116:Vale.AES.AES_common_s.algorithm) = alg in\n va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 569 column 30 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.lemma_gctr_partial_append va_arg116 va_arg117 va_arg118\n va_arg119 va_arg120 va_arg121 va_arg122 va_arg123 va_arg124 va_arg125) (let\n (va_arg115:Vale.Def.Types_s.quad32) = Vale.AES.GCTR.inc32lite (Vale.AES.GCTR.inc32lite ctr_BE_2\n len128x6) len128 in let (va_arg114:Vale.Def.Types_s.quad32) = ctr_BE_2 in let\n (va_arg113:(FStar.Seq.Base.seq Vale.Def.Types_s.nat32)) = key in let\n (va_arg112:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = cipher_byte_seq in let\n (va_arg111:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = plain_byte_seq in let\n (va_arg110:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s) out128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) out128_b) in let\n (va_arg109:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b) in let (va_arg108:Prims.nat) =\n FStar.Seq.Base.length #quad32 plain_byte_seq in let (va_arg107:Prims.nat) = len128x6 + len128\n in let (va_arg106:Vale.AES.AES_common_s.algorithm) = alg in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 575 column 30 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.lemma_gctr_partial_append va_arg106 va_arg107 va_arg108\n va_arg109 va_arg110 va_arg111 va_arg112 va_arg113 va_arg114 va_arg115) (let\n (va_arg105:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = auth_in in let\n (va_arg104:Vale.Def.Types_s.quad32) = y_0 in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 583 column 23 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GHash.lemma_hash_append2 h_LE va_arg104 y_inout y_final va_arg105\n length_quad32) (let auth_in = FStar.Seq.Base.append #quad32 auth_in (FStar.Seq.Base.create\n #Vale.Def.Types_s.quad32 1 length_quad32) in let (va_arg103:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = auth_in in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 585 column 31 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GHash.ghash_incremental_to_ghash h_LE va_arg103) (va_QEmpty\n (()))))))))))))))))))))))))))))))))))))))))))))))))\n\n\nval va_lemma_Gcm_blocks : va_b0:va_code -> va_s0:va_state -> alg:algorithm -> offset:int ->\n auth_b:buffer128 -> abytes_b:buffer128 -> in128x6_b:buffer128 -> out128x6_b:buffer128 ->\n in128_b:buffer128 -> out128_b:buffer128 -> inout_b:buffer128 -> iv_b:buffer128 ->\n scratch_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_blocks alg offset) va_s0 /\\ va_get_ok va_s0 /\\\n (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in let (h_LE:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.X64.Decls.buffer128_read hkeys_b 2\n (va_get_mem_heaplet 0 va_s0)) in sse_enabled /\\ movbe_enabled /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0)\n Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp\n va_s0 + offset + 8) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 24) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 40) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 56) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet\n 1 va_s0) (va_get_reg64 rRdi va_s0) auth_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0)\n Secret /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 7 va_s0) abytes_ptr abytes_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 2\n va_s0) (va_get_reg64 rR8 va_s0) iv_b 1 (va_get_mem_layout va_s0) Public /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 6 va_s0) in128x6_ptr in128x6_b len128x6\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 6\n va_s0) out128x6_ptr out128x6_b len128x6 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0) in128_ptr in128_b len128\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1\n va_s0) out128_ptr out128_b len128 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) inout_ptr inout_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 3\n va_s0) (va_get_reg64 rRbp va_s0) scratch_b 9 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0) hkeys_b 8\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.buffer_disjoints128 iv_b ([keys_b;\n scratch_b; in128x6_b; out128x6_b; hkeys_b; in128_b; out128_b; inout_b]) /\\\n Vale.X64.Decls.buffer_disjoints128 scratch_b ([keys_b; in128x6_b; out128x6_b; in128_b;\n out128_b; inout_b; hkeys_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128x6_b ([keys_b;\n hkeys_b; in128_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128_b ([keys_b; hkeys_b;\n out128x6_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 inout_b ([keys_b; hkeys_b;\n out128x6_b; out128_b]) /\\ (Vale.X64.Decls.buffers_disjoint128 in128x6_b out128x6_b \\/ in128x6_b\n == out128x6_b) /\\ (Vale.X64.Decls.buffers_disjoint128 in128_b out128_b \\/ in128_b == out128_b)\n /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ in128x6_ptr\n + 16 `op_Multiply` len128x6 < pow2_64 /\\ out128x6_ptr + 16 `op_Multiply` len128x6 < pow2_64 /\\\n in128_ptr + 16 `op_Multiply` len128 < pow2_64 /\\ out128_ptr + 16 `op_Multiply` len128 < pow2_64\n /\\ inout_ptr + 16 < pow2_64 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 auth_b ==\n va_get_reg64 rRdx va_s0 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 abytes_b == 1\n /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128x6_b ==\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out128x6_b /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 out128_b /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128\n in128x6_b == len128x6 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b ==\n len128 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 inout_b == 1 /\\\n plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32 /\\ va_get_reg64 rR9 va_s0 + 32 <\n pow2_64 /\\ Vale.X64.Memory.buffer_addr #Vale.X64.Memory.vuint128 keys_b (va_get_mem_heaplet 0\n va_s0) + 128 < pow2_64 /\\ len128x6 `op_Modulus` 6 == 0 /\\ (len128x6 > 0 ==> len128x6 >= 6) /\\\n 12 + len128x6 + 6 < pow2_32 /\\ (va_mul_nat len128x6 (128 `op_Division` 8) + va_mul_nat len128\n (128 `op_Division` 8) <= plain_num_bytes /\\ plain_num_bytes < va_mul_nat len128x6 (128\n `op_Division` 8) + va_mul_nat len128 (128 `op_Division` 8) + 128 `op_Division` 8) /\\\n (va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) <= va_get_reg64 rRsi va_s0 /\\\n va_get_reg64 rRsi va_s0 < va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) + 128\n `op_Division` 8) /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rRcx va_s0)\n (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 (Vale.AES.AES_s.aes_encrypt_LE alg key\n (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0))))))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in let (h_LE:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.X64.Decls.buffer128_read hkeys_b 2\n (va_get_mem_heaplet 0 va_s0)) in Vale.X64.Decls.modifies_buffer128 out128_b (va_get_mem_heaplet\n 1 va_s0) (va_get_mem_heaplet 1 va_sM) /\\ Vale.X64.Decls.modifies_buffer128 iv_b\n (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2 va_sM) /\\ Vale.X64.Decls.modifies_buffer128\n scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3 va_sM) /\\\n Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM) /\\ Vale.X64.Decls.modifies_buffer128 out128x6_b (va_get_mem_heaplet 6 va_s0)\n (va_get_mem_heaplet 6 va_sM) /\\ plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32\n /\\ (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in let\n (ctr_BE_1:quad32) = iv_BE in let (ctr_BE_2:quad32) = Vale.AES.GCTR_s.inc32 iv_BE 1 in let\n (plain_in:(seq quad32)) = (if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) then FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b) else FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in128_b)) in let (cipher_out:(seq quad32)) = (if (plain_num_bytes\n > (len128x6 + len128) `op_Multiply` 128 `op_Division` 8) then FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM)\n out128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM) inout_b) else\n FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM)\n out128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out128_b)) in let\n (cipher_bound:nat) = (if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128 `op_Division`\n 8) then (len128x6 + len128 + 1) else (len128x6 + len128)) in Vale.AES.GCTR.gctr_partial alg\n cipher_bound plain_in cipher_out key ctr_BE_2 /\\ (let (length_quad:quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.Def.Types_s.insert_nat64\n (Vale.Def.Types_s.insert_nat64 (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0) (8\n `op_Multiply` va_get_reg64 rRsi va_s0) 1) (8 `op_Multiply` plain_num_bytes) 0) in let\n (raw_auth_quads:(seq quad32)) = (if (va_get_reg64 rRsi va_s0 > va_get_reg64 rRdx va_s0\n `op_Multiply` 128 `op_Division` 8) then FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 7 va_s0) abytes_b) else Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0)\n auth_b) in let (auth_input_bytes:(seq nat8)) = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_auth_quads) 0 (va_get_reg64 rRsi va_s0) in let\n (padded_auth_bytes:(seq nat8)) = Vale.AES.GCTR_s.pad_to_128_bits auth_input_bytes in let\n (auth_quad_seq:(seq quad32)) = Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_auth_bytes in let\n (raw_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32 (FStar.Seq.Base.append #quad32\n auth_quad_seq (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b))\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b) in let (total_bytes:nat) =\n FStar.Seq.Base.length #quad32 auth_quad_seq `op_Multiply` 16 + plain_num_bytes in let\n (raw_quad_seq:(seq quad32)) = (if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) then (let (ab:(seq nat8)) = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes (FStar.Seq.Base.append #quad32 raw_quad_seq\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b))) 0 total_bytes in let (pb:(seq\n nat8)) = Vale.AES.GCTR_s.pad_to_128_bits ab in Vale.Def.Types_s.le_bytes_to_seq_quad32 pb) else\n raw_quad_seq) in let (auth_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32 raw_quad_seq\n (FStar.Seq.Base.create #quad32 1 length_quad) in va_get_xmm 8 va_sM ==\n Vale.AES.GCTR_s.gctr_encrypt_block ctr_BE_1 (Vale.AES.GHash_s.ghash_LE h_LE auth_quad_seq) alg\n key 0))) /\\ va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 6 va_sM\n (va_update_mem_heaplet 5 va_sM (va_update_mem_heaplet 3 va_sM (va_update_mem_heaplet 2 va_sM\n (va_update_mem_heaplet 1 va_sM (va_update_xmm 15 va_sM (va_update_xmm 14 va_sM (va_update_xmm\n 13 va_sM (va_update_xmm 12 va_sM (va_update_xmm 11 va_sM (va_update_xmm 10 va_sM (va_update_xmm\n 9 va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5\n va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1\n va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR15 va_sM (va_update_reg64 rR14 va_sM\n (va_update_reg64 rR13 va_sM (va_update_reg64 rR12 va_sM (va_update_reg64 rR11 va_sM\n (va_update_reg64 rR10 va_sM (va_update_reg64 rR9 va_sM (va_update_reg64 rR8 va_sM\n (va_update_reg64 rRbp va_sM (va_update_reg64 rRsi va_sM (va_update_reg64 rRdi va_sM\n (va_update_reg64 rRdx va_sM (va_update_reg64 rRcx va_sM (va_update_reg64 rRbx va_sM\n (va_update_reg64 rRax va_sM (va_update_ok va_sM (va_update_mem va_sM\n va_s0)))))))))))))))))))))))))))))))))))))))))\n[@\"opaque_to_smt\"]\nlet va_lemma_Gcm_blocks va_b0 va_s0 alg offset auth_b abytes_b in128x6_b out128x6_b in128_b\n out128_b inout_b iv_b scratch_b key round_keys keys_b hkeys_b =\n let (va_mods:va_mods_t) = [va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5;\n va_Mod_mem_heaplet 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14;\n va_Mod_xmm 13; va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8;\n va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm\n 1; va_Mod_xmm 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12;\n va_Mod_reg64 rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp;\n va_Mod_reg64 rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx;\n va_Mod_reg64 rRax; va_Mod_ok; va_Mod_mem] in\n let va_qc = va_qcode_Gcm_blocks va_mods alg offset auth_b abytes_b in128x6_b out128x6_b in128_b\n out128_b inout_b iv_b scratch_b key round_keys keys_b hkeys_b in\n let (va_sM, va_fM, va_g) = va_wp_sound_code_norm (va_code_Gcm_blocks alg offset) va_qc va_s0 (fun\n va_s0 va_sM va_g -> let () = va_g in label va_range1\n \"***** POSTCONDITION NOT MET AT line 283 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\ (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64\n (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0) in let\n (in128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 8) (va_get_stack va_s0) in let (out128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack va_s0) in\n let (len128x6:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 24) (va_get_stack va_s0) in let (in128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack va_s0) in\n let (out128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 40) (va_get_stack va_s0) in let (len128:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack va_s0) in\n let (inout_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 56) (va_get_stack va_s0) in let (plain_num_bytes:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack va_s0) in\n let (h_LE:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.X64.Decls.buffer128_read hkeys_b 2 (va_get_mem_heaplet 0 va_s0)) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 396 column 56 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out128_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 397 column 52 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 iv_b (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 398 column 57 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 399 column 55 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 400 column 58 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out128x6_b (va_get_mem_heaplet 6 va_s0) (va_get_mem_heaplet\n 6 va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 403 column 39 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (plain_num_bytes < pow2_32) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 404 column 38 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRsi va_s0 < pow2_32) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 406 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 408 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (ctr_BE_1:quad32) = iv_BE in label va_range1\n \"***** POSTCONDITION NOT MET AT line 409 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (ctr_BE_2:quad32) = Vale.AES.GCTR_s.inc32 iv_BE 1 in label va_range1\n \"***** POSTCONDITION NOT MET AT line 412 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (plain_in:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b)) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in128_b)) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 421 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (cipher_out:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM) out128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b)) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_sM) out128_b)) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 430 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (cipher_bound:nat) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> len128x6 + len128 + 1) (fun _ -> len128x6 + len128) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 434 column 77 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.AES.GCTR.gctr_partial alg cipher_bound plain_in cipher_out key ctr_BE_2) /\\ label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 438 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (length_quad:quad32) = Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.Def.Types_s.insert_nat64 (Vale.Def.Types_s.insert_nat64 (Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 0 0 0 0) (8 `op_Multiply` va_get_reg64 rRsi va_s0) 1) (8 `op_Multiply`\n plain_num_bytes) 0) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 440 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (raw_auth_quads:(seq quad32)) = va_if (va_get_reg64 rRsi va_s0 > va_get_reg64 rRdx va_s0\n `op_Multiply` 128 `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 7 va_s0) abytes_b)) (fun _ -> Vale.X64.Decls.s128 (va_get_mem_heaplet 1\n va_s0) auth_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 444 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (auth_input_bytes:(seq nat8)) = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_auth_quads) 0 (va_get_reg64 rRsi va_s0) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 445 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (padded_auth_bytes:(seq nat8)) = Vale.AES.GCTR_s.pad_to_128_bits auth_input_bytes in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 446 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (auth_quad_seq:(seq quad32)) = Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_auth_bytes\n in label va_range1\n \"***** POSTCONDITION NOT MET AT line 448 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (raw_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32 (FStar.Seq.Base.append #quad32\n auth_quad_seq (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b))\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 452 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (total_bytes:nat) = FStar.Seq.Base.length #quad32 auth_quad_seq `op_Multiply` 16 +\n plain_num_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 453 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (raw_quad_seq:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply`\n 128 `op_Division` 8) (fun _ -> let (ab:(seq nat8)) = FStar.Seq.Base.slice\n #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes (FStar.Seq.Base.append #quad32\n raw_quad_seq (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b))) 0 total_bytes in let\n (pb:(seq nat8)) = Vale.AES.GCTR_s.pad_to_128_bits ab in Vale.Def.Types_s.le_bytes_to_seq_quad32\n pb) (fun _ -> raw_quad_seq) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 460 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (auth_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32 raw_quad_seq\n (FStar.Seq.Base.create #quad32 1 length_quad) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 461 column 106 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_xmm 8 va_sM == Vale.AES.GCTR_s.gctr_encrypt_block ctr_BE_1 (Vale.AES.GHash_s.ghash_LE\n h_LE auth_quad_seq) alg key 0)))))))))))))))))) in\n assert_norm (va_qc.mods == va_mods);\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet\n 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13;\n va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm\n 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64\n rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64\n rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64\n rRax; va_Mod_ok; va_Mod_mem]) va_sM va_s0;\n (va_sM, va_fM)\n\n[@ va_qattr]\nlet va_wp_Gcm_blocks (alg:algorithm) (offset:int) (auth_b:buffer128) (abytes_b:buffer128)\n (in128x6_b:buffer128) (out128x6_b:buffer128) (in128_b:buffer128) (out128_b:buffer128)\n (inout_b:buffer128) (iv_b:buffer128) (scratch_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (va_s0:va_state) (va_k:(va_state -> unit ->\n Type0)) : Type0 =\n (va_get_ok va_s0 /\\ (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64\n (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0) in let\n (in128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 8) (va_get_stack va_s0) in let (out128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack va_s0) in\n let (len128x6:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 24) (va_get_stack va_s0) in let (in128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack va_s0) in\n let (out128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 40) (va_get_stack va_s0) in let (len128:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack va_s0) in\n let (inout_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 56) (va_get_stack va_s0) in let (plain_num_bytes:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack va_s0) in\n let (h_LE:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.X64.Decls.buffer128_read hkeys_b 2 (va_get_mem_heaplet 0 va_s0)) in sse_enabled /\\\n movbe_enabled /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 0)\n (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64\n (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 24) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 40) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 56) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet\n 1 va_s0) (va_get_reg64 rRdi va_s0) auth_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0)\n Secret /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 7 va_s0) abytes_ptr abytes_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 2\n va_s0) (va_get_reg64 rR8 va_s0) iv_b 1 (va_get_mem_layout va_s0) Public /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 6 va_s0) in128x6_ptr in128x6_b len128x6\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 6\n va_s0) out128x6_ptr out128x6_b len128x6 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0) in128_ptr in128_b len128\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1\n va_s0) out128_ptr out128_b len128 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) inout_ptr inout_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 3\n va_s0) (va_get_reg64 rRbp va_s0) scratch_b 9 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0) hkeys_b 8\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.buffer_disjoints128 iv_b ([keys_b;\n scratch_b; in128x6_b; out128x6_b; hkeys_b; in128_b; out128_b; inout_b]) /\\\n Vale.X64.Decls.buffer_disjoints128 scratch_b ([keys_b; in128x6_b; out128x6_b; in128_b;\n out128_b; inout_b; hkeys_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128x6_b ([keys_b;\n hkeys_b; in128_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128_b ([keys_b; hkeys_b;\n out128x6_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 inout_b ([keys_b; hkeys_b;\n out128x6_b; out128_b]) /\\ (Vale.X64.Decls.buffers_disjoint128 in128x6_b out128x6_b \\/ in128x6_b\n == out128x6_b) /\\ (Vale.X64.Decls.buffers_disjoint128 in128_b out128_b \\/ in128_b == out128_b)\n /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ in128x6_ptr\n + 16 `op_Multiply` len128x6 < pow2_64 /\\ out128x6_ptr + 16 `op_Multiply` len128x6 < pow2_64 /\\\n in128_ptr + 16 `op_Multiply` len128 < pow2_64 /\\ out128_ptr + 16 `op_Multiply` len128 < pow2_64\n /\\ inout_ptr + 16 < pow2_64 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 auth_b ==\n va_get_reg64 rRdx va_s0 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 abytes_b == 1\n /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128x6_b ==\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out128x6_b /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 out128_b /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128\n in128x6_b == len128x6 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b ==\n len128 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 inout_b == 1 /\\\n plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32 /\\ va_get_reg64 rR9 va_s0 + 32 <\n pow2_64 /\\ Vale.X64.Memory.buffer_addr #Vale.X64.Memory.vuint128 keys_b (va_get_mem_heaplet 0\n va_s0) + 128 < pow2_64 /\\ len128x6 `op_Modulus` 6 == 0 /\\ (len128x6 > 0 ==> len128x6 >= 6) /\\\n 12 + len128x6 + 6 < pow2_32 /\\ (va_mul_nat len128x6 (128 `op_Division` 8) + va_mul_nat len128\n (128 `op_Division` 8) <= plain_num_bytes /\\ plain_num_bytes < va_mul_nat len128x6 (128\n `op_Division` 8) + va_mul_nat len128 (128 `op_Division` 8) + 128 `op_Division` 8) /\\\n (va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) <= va_get_reg64 rRsi va_s0 /\\\n va_get_reg64 rRsi va_s0 < va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) + 128\n `op_Division` 8) /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rRcx va_s0)\n (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 (Vale.AES.AES_s.aes_encrypt_LE alg key\n (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0)))) /\\ (forall (va_x_mem:vale_heap)\n (va_x_rax:nat64) (va_x_rbx:nat64) (va_x_rcx:nat64) (va_x_rdx:nat64) (va_x_rdi:nat64)\n (va_x_rsi:nat64) (va_x_rbp:nat64) (va_x_r8:nat64) (va_x_r9:nat64) (va_x_r10:nat64)\n (va_x_r11:nat64) (va_x_r12:nat64) (va_x_r13:nat64) (va_x_r14:nat64) (va_x_r15:nat64)\n (va_x_xmm0:quad32) (va_x_xmm1:quad32) (va_x_xmm2:quad32) (va_x_xmm3:quad32) (va_x_xmm4:quad32)\n (va_x_xmm5:quad32) (va_x_xmm6:quad32) (va_x_xmm7:quad32) (va_x_xmm8:quad32) (va_x_xmm9:quad32)\n (va_x_xmm10:quad32) (va_x_xmm11:quad32) (va_x_xmm12:quad32) (va_x_xmm13:quad32)\n (va_x_xmm14:quad32) (va_x_xmm15:quad32) (va_x_heap1:vale_heap) (va_x_heap2:vale_heap)\n (va_x_heap3:vale_heap) (va_x_heap5:vale_heap) (va_x_heap6:vale_heap)\n (va_x_efl:Vale.X64.Flags.t) . let va_sM = va_upd_flags va_x_efl (va_upd_mem_heaplet 6\n va_x_heap6 (va_upd_mem_heaplet 5 va_x_heap5 (va_upd_mem_heaplet 3 va_x_heap3\n (va_upd_mem_heaplet 2 va_x_heap2 (va_upd_mem_heaplet 1 va_x_heap1 (va_upd_xmm 15 va_x_xmm15\n (va_upd_xmm 14 va_x_xmm14 (va_upd_xmm 13 va_x_xmm13 (va_upd_xmm 12 va_x_xmm12 (va_upd_xmm 11\n va_x_xmm11 (va_upd_xmm 10 va_x_xmm10 (va_upd_xmm 9 va_x_xmm9 (va_upd_xmm 8 va_x_xmm8\n (va_upd_xmm 7 va_x_xmm7 (va_upd_xmm 6 va_x_xmm6 (va_upd_xmm 5 va_x_xmm5 (va_upd_xmm 4 va_x_xmm4\n (va_upd_xmm 3 va_x_xmm3 (va_upd_xmm 2 va_x_xmm2 (va_upd_xmm 1 va_x_xmm1 (va_upd_xmm 0 va_x_xmm0\n (va_upd_reg64 rR15 va_x_r15 (va_upd_reg64 rR14 va_x_r14 (va_upd_reg64 rR13 va_x_r13\n (va_upd_reg64 rR12 va_x_r12 (va_upd_reg64 rR11 va_x_r11 (va_upd_reg64 rR10 va_x_r10\n (va_upd_reg64 rR9 va_x_r9 (va_upd_reg64 rR8 va_x_r8 (va_upd_reg64 rRbp va_x_rbp (va_upd_reg64\n rRsi va_x_rsi (va_upd_reg64 rRdi va_x_rdi (va_upd_reg64 rRdx va_x_rdx (va_upd_reg64 rRcx\n va_x_rcx (va_upd_reg64 rRbx va_x_rbx (va_upd_reg64 rRax va_x_rax (va_upd_mem va_x_mem\n va_s0))))))))))))))))))))))))))))))))))))) in va_get_ok va_sM /\\ (let\n (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in let (h_LE:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.X64.Decls.buffer128_read hkeys_b 2\n (va_get_mem_heaplet 0 va_s0)) in Vale.X64.Decls.modifies_buffer128 out128_b (va_get_mem_heaplet\n 1 va_s0) (va_get_mem_heaplet 1 va_sM) /\\ Vale.X64.Decls.modifies_buffer128 iv_b\n (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2 va_sM) /\\ Vale.X64.Decls.modifies_buffer128\n scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3 va_sM) /\\\n Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM) /\\ Vale.X64.Decls.modifies_buffer128 out128x6_b (va_get_mem_heaplet 6 va_s0)\n (va_get_mem_heaplet 6 va_sM) /\\ plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32\n /\\ (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in let\n (ctr_BE_1:quad32) = iv_BE in let (ctr_BE_2:quad32) = Vale.AES.GCTR_s.inc32 iv_BE 1 in let\n (plain_in:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b)) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in128_b)) in let (cipher_out:(seq quad32)) = va_if\n (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128 `op_Division` 8) (fun _ ->\n FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_sM) out128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM)\n inout_b)) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM)\n out128_b)) in let (cipher_bound:nat) = va_if (plain_num_bytes > (len128x6 + len128)\n `op_Multiply` 128 `op_Division` 8) (fun _ -> len128x6 + len128 + 1) (fun _ -> len128x6 +\n len128) in Vale.AES.GCTR.gctr_partial alg cipher_bound plain_in cipher_out key ctr_BE_2 /\\ (let\n (length_quad:quad32) = Vale.Def.Types_s.reverse_bytes_quad32 (Vale.Def.Types_s.insert_nat64\n (Vale.Def.Types_s.insert_nat64 (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0) (8\n `op_Multiply` va_get_reg64 rRsi va_s0) 1) (8 `op_Multiply` plain_num_bytes) 0) in let\n (raw_auth_quads:(seq quad32)) = va_if (va_get_reg64 rRsi va_s0 > va_get_reg64 rRdx va_s0\n `op_Multiply` 128 `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 7 va_s0) abytes_b)) (fun _ -> Vale.X64.Decls.s128 (va_get_mem_heaplet 1\n va_s0) auth_b) in let (auth_input_bytes:(seq nat8)) = FStar.Seq.Base.slice\n #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_auth_quads) 0 (va_get_reg64\n rRsi va_s0) in let (padded_auth_bytes:(seq nat8)) = Vale.AES.GCTR_s.pad_to_128_bits\n auth_input_bytes in let (auth_quad_seq:(seq quad32)) = Vale.Def.Types_s.le_bytes_to_seq_quad32\n padded_auth_bytes in let (raw_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32\n (FStar.Seq.Base.append #quad32 auth_quad_seq (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0)\n in128x6_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b) in let (total_bytes:nat)\n = FStar.Seq.Base.length #quad32 auth_quad_seq `op_Multiply` 16 + plain_num_bytes in let\n (raw_quad_seq:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> let (ab:(seq nat8)) = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes (FStar.Seq.Base.append #quad32 raw_quad_seq\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b))) 0 total_bytes in let (pb:(seq\n nat8)) = Vale.AES.GCTR_s.pad_to_128_bits ab in Vale.Def.Types_s.le_bytes_to_seq_quad32 pb) (fun\n _ -> raw_quad_seq) in let (auth_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32\n raw_quad_seq (FStar.Seq.Base.create #quad32 1 length_quad) in va_get_xmm 8 va_sM ==\n Vale.AES.GCTR_s.gctr_encrypt_block ctr_BE_1 (Vale.AES.GHash_s.ghash_LE h_LE auth_quad_seq) alg\n key 0))) ==> va_k va_sM (())))\n\n\nval va_wpProof_Gcm_blocks : alg:algorithm -> offset:int -> auth_b:buffer128 -> abytes_b:buffer128\n -> in128x6_b:buffer128 -> out128x6_b:buffer128 -> in128_b:buffer128 -> out128_b:buffer128 ->\n inout_b:buffer128 -> iv_b:buffer128 -> scratch_b:buffer128 -> key:(seq nat32) -> round_keys:(seq\n quad32) -> keys_b:buffer128 -> hkeys_b:buffer128 -> va_s0:va_state -> va_k:(va_state -> unit ->\n Type0)\n -> Ghost (va_state & va_fuel & unit)\n (requires (va_t_require va_s0 /\\ va_wp_Gcm_blocks alg offset auth_b abytes_b in128x6_b out128x6_b\n in128_b out128_b inout_b iv_b scratch_b key round_keys keys_b hkeys_b va_s0 va_k))\n (ensures (fun (va_sM, va_f0, va_g) -> va_t_ensure (va_code_Gcm_blocks alg offset) ([va_Mod_flags;\n va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet 3; va_Mod_mem_heaplet 2;\n va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13; va_Mod_xmm 12; va_Mod_xmm\n 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5;\n va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR15;\n va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64 rR11; va_Mod_reg64 rR10;\n va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64 rRsi; va_Mod_reg64 rRdi;\n va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64 rRax; va_Mod_mem]) va_s0\n va_k ((va_sM, va_f0, va_g))))\n[@\"opaque_to_smt\"]\nlet va_wpProof_Gcm_blocks alg offset auth_b abytes_b in128x6_b out128x6_b in128_b out128_b inout_b\n iv_b scratch_b key round_keys keys_b hkeys_b va_s0 va_k =\n let (va_sM, va_f0) = va_lemma_Gcm_blocks (va_code_Gcm_blocks alg offset) va_s0 alg offset auth_b\n abytes_b in128x6_b out128x6_b in128_b out128_b inout_b iv_b scratch_b key round_keys keys_b\n hkeys_b in\n va_lemma_upd_update va_sM;\n assert (va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 6 va_sM\n (va_update_mem_heaplet 5 va_sM (va_update_mem_heaplet 3 va_sM (va_update_mem_heaplet 2 va_sM\n (va_update_mem_heaplet 1 va_sM (va_update_xmm 15 va_sM (va_update_xmm 14 va_sM (va_update_xmm\n 13 va_sM (va_update_xmm 12 va_sM (va_update_xmm 11 va_sM (va_update_xmm 10 va_sM (va_update_xmm\n 9 va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5\n va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1\n va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR15 va_sM (va_update_reg64 rR14 va_sM\n (va_update_reg64 rR13 va_sM (va_update_reg64 rR12 va_sM (va_update_reg64 rR11 va_sM\n (va_update_reg64 rR10 va_sM (va_update_reg64 rR9 va_sM (va_update_reg64 rR8 va_sM\n (va_update_reg64 rRbp va_sM (va_update_reg64 rRsi va_sM (va_update_reg64 rRdi va_sM\n (va_update_reg64 rRdx va_sM (va_update_reg64 rRcx va_sM (va_update_reg64 rRbx va_sM\n (va_update_reg64 rRax va_sM (va_update_ok va_sM (va_update_mem va_sM\n va_s0))))))))))))))))))))))))))))))))))))))));\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet\n 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13;\n va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm\n 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64\n rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64\n rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64\n rRax; va_Mod_mem]) va_sM va_s0;\n let va_g = () in\n (va_sM, va_f0, va_g)\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_quick_Gcm_blocks (alg:algorithm) (offset:int) (auth_b:buffer128) (abytes_b:buffer128)\n (in128x6_b:buffer128) (out128x6_b:buffer128) (in128_b:buffer128) (out128_b:buffer128)\n (inout_b:buffer128) (iv_b:buffer128) (scratch_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) : (va_quickCode unit (va_code_Gcm_blocks alg\n offset)) =\n (va_QProc (va_code_Gcm_blocks alg offset) ([va_Mod_flags; va_Mod_mem_heaplet 6;\n va_Mod_mem_heaplet 5; va_Mod_mem_heaplet 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1;\n va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13; va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10;\n va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm\n 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64\n rR13; va_Mod_reg64 rR12; va_Mod_reg64 rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64\n rR8; va_Mod_reg64 rRbp; va_Mod_reg64 rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64\n rRcx; va_Mod_reg64 rRbx; va_Mod_reg64 rRax; va_Mod_mem]) (va_wp_Gcm_blocks alg offset auth_b\n abytes_b in128x6_b out128x6_b in128_b out128_b inout_b iv_b scratch_b key round_keys keys_b\n hkeys_b) (va_wpProof_Gcm_blocks alg offset auth_b abytes_b in128x6_b out128x6_b in128_b\n out128_b inout_b iv_b scratch_b key round_keys keys_b hkeys_b))\n#pop-options\n//--\n//-- Gcm_blocks_wrapped\n#push-options \"--z3rlimit 60\"\n\nval va_code_Gcm_blocks_wrapped : alg:algorithm -> offset:int -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_blocks_wrapped alg offset =\n (va_Block (va_CCons (va_code_Gcm_blocks alg offset) (va_CCons (va_Block (va_CNil ())) (va_CCons\n (va_Block (va_CNil ())) (va_CNil ())))))\n\nval va_codegen_success_Gcm_blocks_wrapped : alg:algorithm -> offset:int -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_blocks_wrapped alg offset =\n (va_pbool_and (va_codegen_success_Gcm_blocks alg offset) (va_ttrue ()))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_blocks_wrapped (va_mods:va_mods_t) (alg:algorithm) (offset:int) (auth_b:buffer128)\n (abytes_b:buffer128) (in128x6_b:buffer128) (out128x6_b:buffer128) (in128_b:buffer128)\n (out128_b:buffer128) (inout_b:buffer128) (iv_b:buffer128) (iv:supported_iv_LE)\n (scratch_b:buffer128) (key:(seq nat32)) (round_keys:(seq quad32)) (keys_b:buffer128)\n (hkeys_b:buffer128) (expected_tag:(seq nat8)) : (va_quickCode unit (va_code_Gcm_blocks_wrapped\n alg offset)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in let\n (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 0) (va_get_stack va_s) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 8) (va_get_stack va_s) in let\n (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 16) (va_get_stack va_s) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 24) (va_get_stack va_s) in let\n (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 32) (va_get_stack va_s) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 40) (va_get_stack va_s) in let\n (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset\n + 48) (va_get_stack va_s) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 56) (va_get_stack va_s) in let\n (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s\n + offset + 64) (va_get_stack va_s) in va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 739 column 15 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_blocks alg offset auth_b abytes_b in128x6_b out128x6_b in128_b out128_b inout_b\n iv_b scratch_b key round_keys keys_b hkeys_b) (fun (va_s:va_state) _ -> va_qAssertSquash\n va_range1\n \"***** EXPRESSION PRECONDITIONS NOT MET WITHIN line 741 column 37 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (8 `op_Multiply` va_get_reg64 rRsi va_old_s >= 0 /\\ 8 `op_Multiply` va_get_reg64 rRsi va_old_s\n <= 18446744073709551615 /\\ 8 `op_Multiply` plain_num_bytes >= 0 /\\ 8 `op_Multiply`\n plain_num_bytes <= 18446744073709551615) (fun _ -> let (va_arg55:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.Def.Types_s.insert_nat64\n (Vale.Def.Types_s.insert_nat64 (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0) (8\n `op_Multiply` va_get_reg64 rRsi va_old_s) 1) (8 `op_Multiply` plain_num_bytes) 0) in let\n (va_arg54:Vale.Def.Types_s.quad32) = va_get_xmm 8 va_s in let\n (va_arg53:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.X64.Decls.buffer128_read hkeys_b 2 (va_get_mem_heaplet 0 va_s)) in let\n (va_arg52:Vale.Def.Types_s.quad32) = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2\n va_old_s) in let (va_arg51:Vale.AES.GCM_s.supported_iv_LE) = iv in let (va_arg50:Prims.nat) =\n va_get_reg64 rRsi va_old_s in let (va_arg49:Prims.nat) = plain_num_bytes in let\n (va_arg48:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s) inout_b in let (va_arg47:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) out128_b in let\n (va_arg46:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_s) out128x6_b in let (va_arg45:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_old_s) inout_b in let\n (va_arg44:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_old_s) in128_b in let (va_arg43:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b in\n let (va_arg42:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 7 va_old_s) abytes_b in let (va_arg41:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) auth_b in let\n (va_arg40:(FStar.Seq.Base.seq Vale.Def.Words_s.nat32)) = key in let\n (va_arg39:Vale.AES.AES_common_s.algorithm) = alg in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 741 column 37 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCM.gcm_blocks_dec_helper_simplified va_arg39 va_arg40 va_arg41\n va_arg42 va_arg43 va_arg44 va_arg45 va_arg46 va_arg47 va_arg48 va_arg49 va_arg50 va_arg51\n va_arg52 va_arg53 va_arg54 va_arg55) (let (auth_raw_quads:(FStar.Seq.Base.seq\n Vale.X64.Decls.quad32)) = FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_old_s) auth_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 7 va_old_s)\n abytes_b) in va_qAssertSquash va_range1\n \"***** EXPRESSION PRECONDITIONS NOT MET WITHIN line 751 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n ((fun a_1906 (s_1907:(FStar.Seq.Base.seq a_1906)) (i_1908:Prims.nat) (j_1909:Prims.nat) -> let\n (j_1869:Prims.nat) = j_1909 in Prims.b2t (Prims.op_AmpAmp (Prims.op_LessThanOrEqual i_1908\n j_1869) (Prims.op_LessThanOrEqual j_1869 (FStar.Seq.Base.length #a_1906 s_1907))))\n Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes auth_raw_quads) 0 (va_get_reg64\n rRsi va_old_s)) (fun _ -> let (auth_bytes:(FStar.Seq.Base.seq Vale.Def.Types_s.nat8)) =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n auth_raw_quads) 0 (va_get_reg64 rRsi va_old_s) in let (va_arg38:Vale.Def.Types_s.quad32) =\n Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_old_s) in let\n (va_arg37:Vale.AES.GCM_s.supported_iv_LE) = iv in let (va_arg36:Prims.nat) = plain_num_bytes in\n let (va_arg35:(FStar.Seq.Base.seq Vale.Def.Words_s.nat8)) = expected_tag in let\n (va_arg34:(FStar.Seq.Base.seq Vale.Def.Words_s.nat8)) = auth_bytes in let\n (va_arg33:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s) inout_b in let (va_arg32:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) out128_b in let\n (va_arg31:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_s) out128x6_b in let (va_arg30:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_old_s) inout_b in let\n (va_arg29:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_old_s) in128_b in let (va_arg28:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b in\n let (va_arg27:(FStar.Seq.Base.seq Vale.Def.Words_s.nat32)) = key in let\n (va_arg26:Vale.AES.AES_common_s.algorithm) = alg in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 752 column 37 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCM.gcm_blocks_helper_dec_simplified va_arg26 va_arg27 va_arg28\n va_arg29 va_arg30 va_arg31 va_arg32 va_arg33 va_arg34 va_arg35 va_arg36 va_arg37 va_arg38)\n (va_QEmpty (()))))))))\n\n\nval va_lemma_Gcm_blocks_wrapped : va_b0:va_code -> va_s0:va_state -> alg:algorithm -> offset:int ->\n auth_b:buffer128 -> abytes_b:buffer128 -> in128x6_b:buffer128 -> out128x6_b:buffer128 ->\n in128_b:buffer128 -> out128_b:buffer128 -> inout_b:buffer128 -> iv_b:buffer128 ->\n iv:supported_iv_LE -> scratch_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) ->\n keys_b:buffer128 -> hkeys_b:buffer128 -> expected_tag:(seq nat8)\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_blocks_wrapped alg offset) va_s0 /\\ va_get_ok\n va_s0 /\\ (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64\n rRsp va_s0 + offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in sse_enabled /\\ movbe_enabled /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0)\n Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp\n va_s0 + offset + 8) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 24) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 40) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 56) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet\n 1 va_s0) (va_get_reg64 rRdi va_s0) auth_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0)\n Secret /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 7 va_s0) abytes_ptr abytes_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 2\n va_s0) (va_get_reg64 rR8 va_s0) iv_b 1 (va_get_mem_layout va_s0) Public /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 6 va_s0) in128x6_ptr in128x6_b len128x6\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 6\n va_s0) out128x6_ptr out128x6_b len128x6 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0) in128_ptr in128_b len128\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1\n va_s0) out128_ptr out128_b len128 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) inout_ptr inout_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 3\n va_s0) (va_get_reg64 rRbp va_s0) scratch_b 9 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0) hkeys_b 8\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.buffer_disjoints128 iv_b ([keys_b;\n scratch_b; in128x6_b; out128x6_b; hkeys_b; in128_b; out128_b; inout_b]) /\\\n Vale.X64.Decls.buffer_disjoints128 scratch_b ([keys_b; in128x6_b; out128x6_b; in128_b;\n out128_b; inout_b; hkeys_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128x6_b ([keys_b;\n hkeys_b; in128_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128_b ([keys_b; hkeys_b;\n out128x6_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 inout_b ([keys_b; hkeys_b;\n out128x6_b; out128_b]) /\\ (Vale.X64.Decls.buffers_disjoint128 in128x6_b out128x6_b \\/ in128x6_b\n == out128x6_b) /\\ (Vale.X64.Decls.buffers_disjoint128 in128_b out128_b \\/ in128_b == out128_b)\n /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ in128x6_ptr\n + 16 `op_Multiply` len128x6 < pow2_64 /\\ out128x6_ptr + 16 `op_Multiply` len128x6 < pow2_64 /\\\n in128_ptr + 16 `op_Multiply` len128 < pow2_64 /\\ out128_ptr + 16 `op_Multiply` len128 < pow2_64\n /\\ inout_ptr + 16 < pow2_64 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 auth_b ==\n va_get_reg64 rRdx va_s0 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 abytes_b == 1\n /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128x6_b ==\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out128x6_b /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 out128_b /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128\n in128x6_b == len128x6 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b ==\n len128 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 inout_b == 1 /\\\n plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32 /\\ va_get_reg64 rR9 va_s0 + 32 <\n pow2_64 /\\ Vale.X64.Memory.buffer_addr #Vale.X64.Memory.vuint128 keys_b (va_get_mem_heaplet 0\n va_s0) + 128 < pow2_64 /\\ len128x6 `op_Modulus` 6 == 0 /\\ (len128x6 > 0 ==> len128x6 >= 6) /\\\n 12 + len128x6 + 6 < pow2_32 /\\ (va_mul_nat len128x6 (128 `op_Division` 8) + va_mul_nat len128\n (128 `op_Division` 8) <= plain_num_bytes /\\ plain_num_bytes < va_mul_nat len128x6 (128\n `op_Division` 8) + va_mul_nat len128 (128 `op_Division` 8) + 128 `op_Division` 8) /\\\n (va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) <= va_get_reg64 rRsi va_s0 /\\\n va_get_reg64 rRsi va_s0 < va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) + 128\n `op_Division` 8) /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rRcx va_s0)\n (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 (Vale.AES.AES_s.aes_encrypt_LE alg key\n (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0))) /\\ (let iv_BE =\n Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in let h_LE =\n Vale.AES.AES_s.aes_encrypt_LE alg key (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0)\n in iv_BE == Vale.AES.GCM_s.compute_iv_BE h_LE iv))))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in Vale.X64.Decls.modifies_buffer128 out128_b\n (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1 va_sM) /\\ Vale.X64.Decls.modifies_buffer128\n iv_b (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2 va_sM) /\\\n Vale.X64.Decls.modifies_buffer128 scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3\n va_sM) /\\ Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0)\n (va_get_mem_heaplet 5 va_sM) /\\ Vale.X64.Decls.modifies_buffer128 out128x6_b\n (va_get_mem_heaplet 6 va_s0) (va_get_mem_heaplet 6 va_sM) /\\ plain_num_bytes < pow2_32 /\\\n va_get_reg64 rRsi va_s0 < pow2_32 /\\ (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0\n (va_get_mem_heaplet 2 va_s0) in let auth_raw_quads = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 7 va_s0) abytes_b) in let auth_bytes =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n auth_raw_quads) 0 (va_get_reg64 rRsi va_s0) in let plain_raw_quads = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_s0) in128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0)\n in128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b) in let plain_bytes =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n plain_raw_quads) 0 plain_num_bytes in let cipher_raw_quads = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM)\n out128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM) inout_b) in let cipher_bytes =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n cipher_raw_quads) 0 plain_num_bytes in l_and (l_and (l_and (FStar.Seq.Base.length\n #Vale.Def.Types_s.nat8 auth_bytes < pow2_32) (FStar.Seq.Base.length #Vale.Def.Types_s.nat8\n plain_bytes < pow2_32)) (cipher_bytes == __proj__Mktuple2__item___1 #(FStar.Seq.Base.seq\n Vale.Def.Types_s.nat8) #bool (Vale.AES.GCM_s.gcm_decrypt_LE alg\n (Vale.Def.Words.Seq_s.seq_nat32_to_seq_nat8_LE key) iv plain_bytes auth_bytes expected_tag)))\n (Vale.Def.Types_s.le_quad32_to_bytes (va_get_xmm 8 va_sM) == Vale.AES.GCM.gcm_decrypt_LE_tag\n alg (Vale.Def.Words.Seq_s.seq_nat32_to_seq_nat8_LE key) iv plain_bytes auth_bytes))) /\\\n va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 6 va_sM (va_update_mem_heaplet\n 5 va_sM (va_update_mem_heaplet 3 va_sM (va_update_mem_heaplet 2 va_sM (va_update_mem_heaplet 1\n va_sM (va_update_xmm 15 va_sM (va_update_xmm 14 va_sM (va_update_xmm 13 va_sM (va_update_xmm 12\n va_sM (va_update_xmm 11 va_sM (va_update_xmm 10 va_sM (va_update_xmm 9 va_sM (va_update_xmm 8\n va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5 va_sM (va_update_xmm 4\n va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1 va_sM (va_update_xmm 0\n va_sM (va_update_reg64 rR15 va_sM (va_update_reg64 rR14 va_sM (va_update_reg64 rR13 va_sM\n (va_update_reg64 rR12 va_sM (va_update_reg64 rR11 va_sM (va_update_reg64 rR10 va_sM\n (va_update_reg64 rR9 va_sM (va_update_reg64 rR8 va_sM (va_update_reg64 rRbp va_sM\n (va_update_reg64 rRsi va_sM (va_update_reg64 rRdi va_sM (va_update_reg64 rRdx va_sM\n (va_update_reg64 rRcx va_sM (va_update_reg64 rRbx va_sM (va_update_reg64 rRax va_sM\n (va_update_ok va_sM (va_update_mem va_sM va_s0)))))))))))))))))))))))))))))))))))))))))\n[@\"opaque_to_smt\"]\nlet va_lemma_Gcm_blocks_wrapped va_b0 va_s0 alg offset auth_b abytes_b in128x6_b out128x6_b in128_b\n out128_b inout_b iv_b iv scratch_b key round_keys keys_b hkeys_b expected_tag =\n let (va_mods:va_mods_t) = [va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5;\n va_Mod_mem_heaplet 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14;\n va_Mod_xmm 13; va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8;\n va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm\n 1; va_Mod_xmm 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12;\n va_Mod_reg64 rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp;\n va_Mod_reg64 rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx;\n va_Mod_reg64 rRax; va_Mod_ok; va_Mod_mem] in\n let va_qc = va_qcode_Gcm_blocks_wrapped va_mods alg offset auth_b abytes_b in128x6_b out128x6_b\n in128_b out128_b inout_b iv_b iv scratch_b key round_keys keys_b hkeys_b expected_tag in\n let (va_sM, va_fM, va_g) = va_wp_sound_code_norm (va_code_Gcm_blocks_wrapped alg offset) va_qc\n va_s0 (fun va_s0 va_sM va_g -> let () = va_g in label va_range1\n \"***** POSTCONDITION NOT MET AT line 588 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\ (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64\n (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0) in let\n (in128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 8) (va_get_stack va_s0) in let (out128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack va_s0) in\n let (len128x6:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 24) (va_get_stack va_s0) in let (in128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack va_s0) in\n let (out128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 40) (va_get_stack va_s0) in let (len128:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack va_s0) in\n let (inout_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 56) (va_get_stack va_s0) in let (plain_num_bytes:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack va_s0) in\n label va_range1\n \"***** POSTCONDITION NOT MET AT line 711 column 56 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out128_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 712 column 52 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 iv_b (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 713 column 57 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 714 column 55 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 715 column 58 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out128x6_b (va_get_mem_heaplet 6 va_s0) (va_get_mem_heaplet\n 6 va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 718 column 39 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (plain_num_bytes < pow2_32) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 719 column 38 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRsi va_s0 < pow2_32) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 721 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 723 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let auth_raw_quads = FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) auth_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 7 va_s0)\n abytes_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 724 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let auth_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes auth_raw_quads) 0 (va_get_reg64 rRsi va_s0) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 725 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let plain_raw_quads = FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 726 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let plain_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes plain_raw_quads) 0 plain_num_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 727 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let cipher_raw_quads = FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM) out128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 728 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let cipher_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes cipher_raw_quads) 0 plain_num_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 737 column 52 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (l_and (l_and (l_and (FStar.Seq.Base.length #Vale.Def.Types_s.nat8 auth_bytes < pow2_32)\n (FStar.Seq.Base.length #Vale.Def.Types_s.nat8 plain_bytes < pow2_32)) (cipher_bytes ==\n __proj__Mktuple2__item___1 #(FStar.Seq.Base.seq Vale.Def.Types_s.nat8) #bool\n (Vale.AES.GCM_s.gcm_decrypt_LE alg (Vale.Def.Words.Seq_s.seq_nat32_to_seq_nat8_LE key) iv\n plain_bytes auth_bytes expected_tag))) (Vale.Def.Types_s.le_quad32_to_bytes (va_get_xmm 8\n va_sM) == Vale.AES.GCM.gcm_decrypt_LE_tag alg (Vale.Def.Words.Seq_s.seq_nat32_to_seq_nat8_LE\n key) iv plain_bytes auth_bytes))))))))))) in\n assert_norm (va_qc.mods == va_mods);\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet\n 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13;\n va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm\n 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64\n rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64\n rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64\n rRax; va_Mod_ok; va_Mod_mem]) va_sM va_s0;\n (va_sM, va_fM)\n\n[@ va_qattr]\nlet va_wp_Gcm_blocks_wrapped (alg:algorithm) (offset:int) (auth_b:buffer128) (abytes_b:buffer128)\n (in128x6_b:buffer128) (out128x6_b:buffer128) (in128_b:buffer128) (out128_b:buffer128)\n (inout_b:buffer128) (iv_b:buffer128) (iv:supported_iv_LE) (scratch_b:buffer128) (key:(seq nat32))\n (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (expected_tag:(seq nat8))\n (va_s0:va_state) (va_k:(va_state -> unit -> Type0)) : Type0 =\n (va_get_ok va_s0 /\\ (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64\n (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0) in let\n (in128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 8) (va_get_stack va_s0) in let (out128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack va_s0) in\n let (len128x6:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 24) (va_get_stack va_s0) in let (in128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack va_s0) in\n let (out128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 40) (va_get_stack va_s0) in let (len128:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack va_s0) in\n let (inout_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 56) (va_get_stack va_s0) in let (plain_num_bytes:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack va_s0) in\n sse_enabled /\\ movbe_enabled /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 +\n offset + 0) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0)\n Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 32) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 48) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 64) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0) (va_get_reg64 rRdi va_s0) auth_b\n (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 7 va_s0) abytes_ptr abytes_b 1 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 2 va_s0) (va_get_reg64 rR8 va_s0) iv_b 1\n (va_get_mem_layout va_s0) Public /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 6\n va_s0) in128x6_ptr in128x6_b len128x6 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 6 va_s0) out128x6_ptr out128x6_b len128x6\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1\n va_s0) in128_ptr in128_b len128 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1 va_s0) out128_ptr out128_b len128\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5\n va_s0) inout_ptr inout_b 1 (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128\n (va_get_mem_heaplet 3 va_s0) (va_get_reg64 rRbp va_s0) scratch_b 9 (va_get_mem_layout va_s0)\n Secret /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0)\n hkeys_b 8 (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.buffer_disjoints128 iv_b ([keys_b;\n scratch_b; in128x6_b; out128x6_b; hkeys_b; in128_b; out128_b; inout_b]) /\\\n Vale.X64.Decls.buffer_disjoints128 scratch_b ([keys_b; in128x6_b; out128x6_b; in128_b;\n out128_b; inout_b; hkeys_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128x6_b ([keys_b;\n hkeys_b; in128_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128_b ([keys_b; hkeys_b;\n out128x6_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 inout_b ([keys_b; hkeys_b;\n out128x6_b; out128_b]) /\\ (Vale.X64.Decls.buffers_disjoint128 in128x6_b out128x6_b \\/ in128x6_b\n == out128x6_b) /\\ (Vale.X64.Decls.buffers_disjoint128 in128_b out128_b \\/ in128_b == out128_b)\n /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ in128x6_ptr\n + 16 `op_Multiply` len128x6 < pow2_64 /\\ out128x6_ptr + 16 `op_Multiply` len128x6 < pow2_64 /\\\n in128_ptr + 16 `op_Multiply` len128 < pow2_64 /\\ out128_ptr + 16 `op_Multiply` len128 < pow2_64\n /\\ inout_ptr + 16 < pow2_64 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 auth_b ==\n va_get_reg64 rRdx va_s0 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 abytes_b == 1\n /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128x6_b ==\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out128x6_b /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 out128_b /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128\n in128x6_b == len128x6 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b ==\n len128 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 inout_b == 1 /\\\n plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32 /\\ va_get_reg64 rR9 va_s0 + 32 <\n pow2_64 /\\ Vale.X64.Memory.buffer_addr #Vale.X64.Memory.vuint128 keys_b (va_get_mem_heaplet 0\n va_s0) + 128 < pow2_64 /\\ len128x6 `op_Modulus` 6 == 0 /\\ (len128x6 > 0 ==> len128x6 >= 6) /\\\n 12 + len128x6 + 6 < pow2_32 /\\ (va_mul_nat len128x6 (128 `op_Division` 8) + va_mul_nat len128\n (128 `op_Division` 8) <= plain_num_bytes /\\ plain_num_bytes < va_mul_nat len128x6 (128\n `op_Division` 8) + va_mul_nat len128 (128 `op_Division` 8) + 128 `op_Division` 8) /\\\n (va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) <= va_get_reg64 rRsi va_s0 /\\\n va_get_reg64 rRsi va_s0 < va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) + 128\n `op_Division` 8) /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rRcx va_s0)\n (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 (Vale.AES.AES_s.aes_encrypt_LE alg key\n (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0))) /\\ (let iv_BE =\n Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in let h_LE =\n Vale.AES.AES_s.aes_encrypt_LE alg key (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0)\n in iv_BE == Vale.AES.GCM_s.compute_iv_BE h_LE iv)) /\\ (forall (va_x_mem:vale_heap)\n (va_x_rax:nat64) (va_x_rbx:nat64) (va_x_rcx:nat64) (va_x_rdx:nat64) (va_x_rdi:nat64)\n (va_x_rsi:nat64) (va_x_rbp:nat64) (va_x_r8:nat64) (va_x_r9:nat64) (va_x_r10:nat64)\n (va_x_r11:nat64) (va_x_r12:nat64) (va_x_r13:nat64) (va_x_r14:nat64) (va_x_r15:nat64)\n (va_x_xmm0:quad32) (va_x_xmm1:quad32) (va_x_xmm2:quad32) (va_x_xmm3:quad32) (va_x_xmm4:quad32)\n (va_x_xmm5:quad32) (va_x_xmm6:quad32) (va_x_xmm7:quad32) (va_x_xmm8:quad32) (va_x_xmm9:quad32)\n (va_x_xmm10:quad32) (va_x_xmm11:quad32) (va_x_xmm12:quad32) (va_x_xmm13:quad32)\n (va_x_xmm14:quad32) (va_x_xmm15:quad32) (va_x_heap1:vale_heap) (va_x_heap2:vale_heap)\n (va_x_heap3:vale_heap) (va_x_heap5:vale_heap) (va_x_heap6:vale_heap)\n (va_x_efl:Vale.X64.Flags.t) . let va_sM = va_upd_flags va_x_efl (va_upd_mem_heaplet 6\n va_x_heap6 (va_upd_mem_heaplet 5 va_x_heap5 (va_upd_mem_heaplet 3 va_x_heap3\n (va_upd_mem_heaplet 2 va_x_heap2 (va_upd_mem_heaplet 1 va_x_heap1 (va_upd_xmm 15 va_x_xmm15\n (va_upd_xmm 14 va_x_xmm14 (va_upd_xmm 13 va_x_xmm13 (va_upd_xmm 12 va_x_xmm12 (va_upd_xmm 11\n va_x_xmm11 (va_upd_xmm 10 va_x_xmm10 (va_upd_xmm 9 va_x_xmm9 (va_upd_xmm 8 va_x_xmm8\n (va_upd_xmm 7 va_x_xmm7 (va_upd_xmm 6 va_x_xmm6 (va_upd_xmm 5 va_x_xmm5 (va_upd_xmm 4 va_x_xmm4\n (va_upd_xmm 3 va_x_xmm3 (va_upd_xmm 2 va_x_xmm2 (va_upd_xmm 1 va_x_xmm1 (va_upd_xmm 0 va_x_xmm0\n (va_upd_reg64 rR15 va_x_r15 (va_upd_reg64 rR14 va_x_r14 (va_upd_reg64 rR13 va_x_r13\n (va_upd_reg64 rR12 va_x_r12 (va_upd_reg64 rR11 va_x_r11 (va_upd_reg64 rR10 va_x_r10\n (va_upd_reg64 rR9 va_x_r9 (va_upd_reg64 rR8 va_x_r8 (va_upd_reg64 rRbp va_x_rbp (va_upd_reg64\n rRsi va_x_rsi (va_upd_reg64 rRdi va_x_rdi (va_upd_reg64 rRdx va_x_rdx (va_upd_reg64 rRcx\n va_x_rcx (va_upd_reg64 rRbx va_x_rbx (va_upd_reg64 rRax va_x_rax (va_upd_mem va_x_mem\n va_s0))))))))))))))))))))))))))))))))))))) in va_get_ok va_sM /\\ (let\n (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in Vale.X64.Decls.modifies_buffer128 out128_b\n (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1 va_sM) /\\ Vale.X64.Decls.modifies_buffer128\n iv_b (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2 va_sM) /\\\n Vale.X64.Decls.modifies_buffer128 scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3\n va_sM) /\\ Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0)\n (va_get_mem_heaplet 5 va_sM) /\\ Vale.X64.Decls.modifies_buffer128 out128x6_b\n (va_get_mem_heaplet 6 va_s0) (va_get_mem_heaplet 6 va_sM) /\\ plain_num_bytes < pow2_32 /\\\n va_get_reg64 rRsi va_s0 < pow2_32 /\\ (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0\n (va_get_mem_heaplet 2 va_s0) in let auth_raw_quads = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 7 va_s0) abytes_b) in let auth_bytes =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n auth_raw_quads) 0 (va_get_reg64 rRsi va_s0) in let plain_raw_quads = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_s0) in128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0)\n in128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b) in let plain_bytes =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n plain_raw_quads) 0 plain_num_bytes in let cipher_raw_quads = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM)\n out128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM) inout_b) in let cipher_bytes =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n cipher_raw_quads) 0 plain_num_bytes in l_and (l_and (l_and (FStar.Seq.Base.length\n #Vale.Def.Types_s.nat8 auth_bytes < pow2_32) (FStar.Seq.Base.length #Vale.Def.Types_s.nat8\n plain_bytes < pow2_32)) (cipher_bytes == __proj__Mktuple2__item___1 #(FStar.Seq.Base.seq\n Vale.Def.Types_s.nat8) #bool (Vale.AES.GCM_s.gcm_decrypt_LE alg\n (Vale.Def.Words.Seq_s.seq_nat32_to_seq_nat8_LE key) iv plain_bytes auth_bytes expected_tag)))\n (Vale.Def.Types_s.le_quad32_to_bytes (va_get_xmm 8 va_sM) == Vale.AES.GCM.gcm_decrypt_LE_tag\n alg (Vale.Def.Words.Seq_s.seq_nat32_to_seq_nat8_LE key) iv plain_bytes auth_bytes))) ==> va_k\n va_sM (())))\n\n\nval va_wpProof_Gcm_blocks_wrapped : alg:algorithm -> offset:int -> auth_b:buffer128 ->\n abytes_b:buffer128 -> in128x6_b:buffer128 -> out128x6_b:buffer128 -> in128_b:buffer128 ->\n out128_b:buffer128 -> inout_b:buffer128 -> iv_b:buffer128 -> iv:supported_iv_LE ->\n scratch_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128 -> expected_tag:(seq nat8) -> va_s0:va_state -> va_k:(va_state -> unit -> Type0)\n -> Ghost (va_state & va_fuel & unit)\n (requires (va_t_require va_s0 /\\ va_wp_Gcm_blocks_wrapped alg offset auth_b abytes_b in128x6_b\n out128x6_b in128_b out128_b inout_b iv_b iv scratch_b key round_keys keys_b hkeys_b\n expected_tag va_s0 va_k))\n (ensures (fun (va_sM, va_f0, va_g) -> va_t_ensure (va_code_Gcm_blocks_wrapped alg offset)\n ([va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet 3;\n va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13;\n va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm\n 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64\n rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64\n rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64\n rRax; va_Mod_mem]) va_s0 va_k ((va_sM, va_f0, va_g))))\n\nPartial implementation:\nlet va_wpProof_Gcm_blocks_wrapped\n alg\n offset\n auth_b\n abytes_b\n in128x6_b\n out128x6_b\n in128_b\n out128_b\n inout_b\n iv_b\n iv\n scratch_b\n key\n round_keys\n keys_b\n hkeys_b\n expected_tag\n va_s0\n va_k\n =\n \n"
},
{
"role": "assistant",
"content": "let va_sM, va_f0 =\n va_lemma_Gcm_blocks_wrapped (va_code_Gcm_blocks_wrapped alg offset) va_s0 alg offset auth_b\n abytes_b in128x6_b out128x6_b in128_b out128_b inout_b iv_b iv scratch_b key round_keys keys_b\n hkeys_b expected_tag\nin\nva_lemma_upd_update va_sM;\nassert (va_state_eq va_sM\n (va_update_flags va_sM\n (va_update_mem_heaplet 6\n va_sM\n (va_update_mem_heaplet 5\n va_sM\n (va_update_mem_heaplet 3\n va_sM\n (va_update_mem_heaplet 2\n va_sM\n (va_update_mem_heaplet 1\n va_sM\n (va_update_xmm 15\n va_sM\n (va_update_xmm 14\n va_sM\n (va_update_xmm 13\n va_sM\n (va_update_xmm 12\n va_sM\n (va_update_xmm 11\n va_sM\n (va_update_xmm 10\n va_sM\n (va_update_xmm 9\n va_sM\n (va_update_xmm 8\n va_sM\n (va_update_xmm 7\n va_sM\n (va_update_xmm 6\n va_sM\n (va_update_xmm 5\n va_sM\n (va_update_xmm 4\n va_sM\n (va_update_xmm 3\n va_sM\n (va_update_xmm 2\n va_sM\n (va_update_xmm\n 1\n va_sM\n (va_update_xmm\n 0\n va_sM\n (va_update_reg64\n rR15\n va_sM\n (va_update_reg64\n rR14\n va_sM\n (\n va_update_reg64\n rR13\n va_sM\n (\n va_update_reg64\n rR12\n va_sM\n (\n va_update_reg64\n rR11\n va_sM\n (\n va_update_reg64\n rR10\n va_sM\n (\n va_update_reg64\n rR9\n va_sM\n (\n va_update_reg64\n rR8\n va_sM\n (\n va_update_reg64\n rRbp\n va_sM\n (\n va_update_reg64\n rRsi\n va_sM\n (\n va_update_reg64\n rRdi\n va_sM\n (\n va_update_reg64\n rRdx\n va_sM\n (\n va_update_reg64\n rRcx\n va_sM\n (\n va_update_reg64\n rRbx\n va_sM\n (\n va_update_reg64\n rRax\n va_sM\n (\n va_update_ok\n va_sM\n (\n va_update_mem\n va_sM\n va_s0\n\n )\n\n )\n\n )\n\n )\n\n )\n\n )\n\n )\n\n )\n\n )\n\n )\n\n )\n\n )\n\n )\n\n )\n\n )\n\n ))\n ))))))))))\n )))))))))))));\nva_lemma_norm_mods ([\n va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet 3;\n va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13;\n va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1;\n va_Mod_xmm 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12;\n va_Mod_reg64 rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp;\n va_Mod_reg64 rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx;\n va_Mod_reg64 rRax; va_Mod_mem\n ])\n va_sM\n va_s0;\nlet va_g = () in\n(va_sM, va_f0, va_g)"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n l: LowStar.Lib.LinkedList.t a ->\n n: Prims.list a ->\n r: LowStar.Monotonic.Buffer.loc ->\n h0: FStar.Monotonic.HyperStack.mem ->\n h1: FStar.Monotonic.HyperStack.mem\n -> FStar.Pervasives.Lemma\n (requires\n LowStar.Lib.LinkedList.well_formed h0 l n /\\ LowStar.Lib.LinkedList.invariant h0 l n /\\\n LowStar.Monotonic.Buffer.loc_disjoint r (LowStar.Lib.LinkedList.footprint h0 l n) /\\\n LowStar.Monotonic.Buffer.modifies r h0 h1)\n (ensures\n LowStar.Lib.LinkedList.well_formed h1 l n /\\\n LowStar.Lib.LinkedList.footprint h1 l n == LowStar.Lib.LinkedList.footprint h0 l n /\\\n LowStar.Lib.LinkedList.cells h1 l n == LowStar.Lib.LinkedList.cells h0 l n /\\\n LowStar.Lib.LinkedList.invariant h1 l n)\n (decreases n)\n [\n SMTPat (LowStar.Lib.LinkedList.well_formed h1 l n);\n SMTPat (LowStar.Monotonic.Buffer.modifies r h0 h1)\n ]\nEffect: FStar.Pervasives.Lemma\n\nContext:\nmodule LowStar.Lib.LinkedList\n\n/// This module, oftentimes referred to as LL1 in other parts of the\n/// documentation, provides a low-level view over a linked list. Unless you\n/// intend on modifying the structure of the linked list (e.g. removing cells)\n/// or iterating over each cell, your are probably better off using LL2.\n///\n/// This module is intentionally not relying on a tight abstraction, to i) permit\n/// direct-style iteration by clients and ii) allow stating some SMTPat lemmas\n/// only on LL1 rather than replicating them in LL2 (going from LL2 to LL1 can\n/// happen automatically at the SMT level since LL2 does not have an fsti either).\n///\n/// Some open questions relative to the design of this module:\n/// - should ``v`` be defined directly in terms of ``cells`` rather than have a\n/// recursive definition along with an equivalence lemma?\n/// - should ``footprint`` be defined directly in terms of ``cells`` rather than\n/// have a recursive definition along with an equivalence lemma?\n/// - is it worth the trouble defining a footprint? my intuition is that it's\n/// fine and allows some clients to do very precise reasoning -- other modules\n/// (e.g. AssocList), when establishing abstraction boundaries, can simplify\n/// things and get rid of the precise footprint reasoning in favor of a static\n/// region-based footprint\n///\n/// Some remaining work items:\n/// - switch to C.Loops.while\n\nopen LowStar.BufferOps\n\nmodule B = LowStar.Buffer\nmodule HS = FStar.HyperStack\nmodule G = FStar.Ghost\nmodule L = FStar.List.Tot\nmodule U32 = FStar.UInt32\nmodule ST = FStar.HyperStack.ST\n\nopen FStar.HyperStack.ST\n\n/// Breaking from my own recommendation here... using fuel 1 / ifuel 1 since\n/// everything does recursive list reasoning.\n#set-options \"--__no_positivity --fuel 1 --ifuel 1\"\n\n/// Definition of the Low* type\n/// ---------------------------\nnoeq\ntype t (a: Type0) =\n B.pointer_or_null (cell a)\n\nand cell (a: Type0) = {\n next: t a;\n data: a;\n}\n\n/// Unlike the canonical style that associates a ``v`` function to any stateful\n/// representation, we use a relation here. Writing a function is just\n/// impossible, since we can't prove termination of ``v`` owing to the fact that\n/// there may be an infinite number of cells in the heap.\n///\n/// LL2 is the module that switches to the function style.\n///\n/// Note: no need to use erased here for ``l`` since the function is already in ``GTot``.\nlet rec well_formed #a (h: HS.mem) (c: t a) (l: list a): GTot Type0 (decreases l) =\n B.live h c /\\ (\n match l with\n | [] ->\n B.g_is_null c\n | a :: q ->\n B.length c == 1 /\\ (\n let { data=data; next=next } = B.get h c 0 in\n a == data /\\\n well_formed h next (G.hide q)\n ))\n\n/// Precise reasoning via the cells predicate\n/// -----------------------------------------\n///\n/// The modifies clauses are given either on the footprint of the list, or on a\n/// region that includes the footprint of the list (higher up the stack, e.g.\n/// LL2). This is good, but does not rule out \"stupid\" implementations that\n/// might re-allocate every single list cell when, say, doing a pop.\n///\n/// For advanced usages which take pointers directly to individual list cells,\n/// it's important to rule out these cases. We thus define the ``cells``\n/// predicate for that purpose.\n\nlet rec cells #a (h: HS.mem) (c: t a) (l: list a): Ghost (list (B.pointer (cell a)))\n (requires well_formed h c l)\n (ensures fun _ -> True)\n (decreases l)\n=\n if B.g_is_null c then\n []\n else\n c :: cells h (B.deref h c).next (List.Tot.tl l)\n\n/// A missing bit of automated reasoning\n/// ------------------------------------\n\n/// This should be absolutely trivial, yet, because there's no pattern on\n/// null_unique, we have to do the case analysis and the decomposition\n/// ourselves. This is quite tedious, as it usually takes a while (at least it\n/// did for me!) to realize that this property does not hold automatically.\n///\n/// I'm setting an SMTPat on this lemma because it's absolutely impossible to\n/// get any work done without it.\nlet same_cells_same_pointer #a (h0 h1: HS.mem) (ll0 ll1: t a) (l0 l1: list a): Lemma\n (requires (\n well_formed h0 ll0 l0 /\\\n well_formed h1 ll1 l1 /\\\n cells h0 ll0 l0 == cells h1 ll1 l1))\n (ensures (\n ll0 == ll1))\n [ SMTPat (cells h0 ll0 l0); SMTPat (cells h1 ll1 l1) ]\n=\n if B.g_is_null ll0 && B.g_is_null ll1 then begin\n B.null_unique ll0;\n B.null_unique ll1\n end else if not (B.g_is_null ll0) && not (B.g_is_null ll1) then begin\n ()\n end else\n false_elim ()\n\n/// Classic stateful reasoning lemmas\n/// ---------------------------------\n///\n/// This essentially follows ``stateful``, the type class of a mutable piece of\n/// state defined using the modifies-theory of Low*v2 and equipped with the\n/// right lemmas to enable usage by clients. See\n/// hacl-star/code/streaming/Hacl.Streaming.Interface.fst for more context.\n\nval footprint: (#a: Type) -> (h: HS.mem) -> (l: t a) -> (n: list a) -> Ghost B.loc\n (requires (well_formed h l n))\n (ensures (fun refs -> True))\n (decreases n)\n\nlet rec footprint #a h l n =\n if B.g_is_null l\n then B.loc_none\n else\n let {next = next} = B.get h l 0 in\n let refs = footprint h next (G.hide (L.tl n)) in\n B.loc_union (B.loc_addr_of_buffer l) refs\n\n/// Departing from LinkedList4 here. I prefer to bolt these into the invariant\n/// rather than requiring on some wizard lemmas from LowStar.Monotonic.Buffer to\n/// deduce these.\nval cells_pairwise_disjoint: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_pairwise_disjoint #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.loc_disjoint (B.loc_addr_of_buffer l) (footprint h next (L.tl n)) /\\\n cells_pairwise_disjoint h next (L.tl n)\n\nval cells_live_freeable: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_live_freeable #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.live h l /\\ B.freeable l /\\\n cells_live_freeable h next (L.tl n)\n\nval invariant: #a:Type -> h:HS.mem -> l: t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\nlet invariant #a h l n =\n cells_live_freeable h l n /\\ cells_pairwise_disjoint h l n\n\n/// Normally this would be automatic and writing a custom lemma for that stuff\n/// would be needed only when writing an fsti, however, as stated earlier,\n/// recursion makes everything more difficult and this absolutely needs to be in\n/// scope otherwise the modifies-clause theory cannot deduce disjointness of\n/// fresh allocations w.r.t. the footprint.\nlet rec invariant_loc_in_footprint (#a: Type) (h: HS.mem) (l: t a) (n: list a): Lemma\n (requires well_formed h l n)\n (ensures B.loc_in (footprint h l n) h)\n (decreases n)\n [ SMTPat (invariant h l n) ]\n=\n if B.g_is_null l then\n ()\n else\n let next = (B.deref h l).next in\n invariant_loc_in_footprint h next (L.tl n)\n\n/// Another absolutely essential lemma, for which interestingly the pattern\n/// \"invariant h1 l n\" does not work, but well_formed does... I guess that's\n\nPartial implementation:\nlet rec frame (#a: Type) (l: t a) (n: list a) (r: B.loc) (h0 h1: HS.mem)\n : Lemma\n (requires\n (well_formed h0 l n /\\ invariant h0 l n /\\ B.loc_disjoint r (footprint h0 l n) /\\\n B.modifies r h0 h1))\n (ensures\n (well_formed h1 l n /\\ footprint h1 l n == footprint h0 l n /\\ cells h1 l n == cells h0 l n /\\\n invariant h1 l n))\n (decreases n)\n [SMTPat (well_formed h1 l n); SMTPat (B.modifies r h0 h1)] =\n \n"
},
{
"role": "assistant",
"content": "if B.g_is_null l then () else frame (B.deref h0 l).next (L.tl n) r h0 h1"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n s: LowParse.Spec.Base.serializer p ->\n sz: Prims.nat ->\n sz32: FStar.UInt32.t{FStar.UInt32.v sz32 == sz}\n -> LowParse.SLow.Base.size32 (LowParse.Spec.FLData.serialize_fldata_strong s sz)\nEffect: Prims.Tot\n\nContext:\nmodule LowParse.SLow.FLData\ninclude LowParse.Spec.FLData\ninclude LowParse.SLow.Combinators\n\nmodule B32 = LowParse.Bytes32\nmodule U32 = FStar.UInt32\n\ninline_for_extraction\nlet parse32_fldata\n (#k: parser_kind)\n (#t: Type)\n (#p: parser k t)\n (p32: parser32 p)\n (sz: nat)\n (sz32: U32.t { U32.v sz32 == sz } )\n: Tot (parser32 (parse_fldata p sz))\n= (fun (input: bytes32) -> ((\n if U32.lt (B32.len input) sz32\n then\n None\n else\n match p32 (B32.b32slice input 0ul sz32) with\n | Some (v, consumed) ->\n\tif consumed = sz32\n\tthen begin\n\t Some (v, consumed)\n\tend else None\n | None -> None\n ) <: (res: option (t * U32.t) { parser32_correct (parse_fldata p sz) input res } )))\n\ninline_for_extraction\nlet parse32_fldata_strong\n (#k: parser_kind)\n (#t: Type)\n (#p: parser k t)\n (s: serializer p)\n (p32: parser32 p)\n (sz: nat)\n (sz32: U32.t { U32.v sz32 == sz } )\n: Tot (parser32 (parse_fldata_strong s sz))\n= (fun (input: bytes32) -> ((\n match parse32_fldata p32 sz sz32 input with\n | Some (v, consumed) ->\n assert (\n parse_fldata_strong_correct s sz (B32.reveal input) (U32.v consumed) v;\n Seq.length (s v) == sz\n );\n Some ((v <: parse_fldata_strong_t s sz), consumed)\n | None -> None\n ) \n <: (res: option (parse_fldata_strong_t s sz * U32.t) { parser32_correct (parse_fldata_strong s sz) input res } )))\n\ninline_for_extraction\nlet serialize32_fldata_strong\n (#k: parser_kind)\n (#t: Type)\n (#p: parser k t)\n (#s: serializer p)\n (s32: partial_serializer32 s)\n (sz: nat { sz < 4294967296 } )\n: Tot (serializer32 (serialize_fldata_strong s sz))\n= (fun (input: parse_fldata_strong_t s sz) -> s32 input)\n\nPartial implementation:\nlet size32_fldata_strong\n (#k: parser_kind)\n (#t: Type)\n (#p: parser k t)\n (s: serializer p)\n (sz: nat)\n (sz32: U32.t{U32.v sz32 == sz})\n : Tot (size32 (serialize_fldata_strong s sz)) =\n \n"
},
{
"role": "assistant",
"content": "size32_constant (serialize_fldata_strong s sz) sz32 ()"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n r: FStar.Monotonic.HyperHeap.rid ->\n n: FStar.Ghost.erased (Prims.list a) ->\n pl: LowStar.Buffer.pointer (LowStar.Lib.LinkedList.t a) ->\n x: a\n -> FStar.HyperStack.ST.ST Prims.unit\nEffect: FStar.HyperStack.ST.ST\n\nContext:\nmodule LowStar.Lib.LinkedList\n\n/// This module, oftentimes referred to as LL1 in other parts of the\n/// documentation, provides a low-level view over a linked list. Unless you\n/// intend on modifying the structure of the linked list (e.g. removing cells)\n/// or iterating over each cell, your are probably better off using LL2.\n///\n/// This module is intentionally not relying on a tight abstraction, to i) permit\n/// direct-style iteration by clients and ii) allow stating some SMTPat lemmas\n/// only on LL1 rather than replicating them in LL2 (going from LL2 to LL1 can\n/// happen automatically at the SMT level since LL2 does not have an fsti either).\n///\n/// Some open questions relative to the design of this module:\n/// - should ``v`` be defined directly in terms of ``cells`` rather than have a\n/// recursive definition along with an equivalence lemma?\n/// - should ``footprint`` be defined directly in terms of ``cells`` rather than\n/// have a recursive definition along with an equivalence lemma?\n/// - is it worth the trouble defining a footprint? my intuition is that it's\n/// fine and allows some clients to do very precise reasoning -- other modules\n/// (e.g. AssocList), when establishing abstraction boundaries, can simplify\n/// things and get rid of the precise footprint reasoning in favor of a static\n/// region-based footprint\n///\n/// Some remaining work items:\n/// - switch to C.Loops.while\n\nopen LowStar.BufferOps\n\nmodule B = LowStar.Buffer\nmodule HS = FStar.HyperStack\nmodule G = FStar.Ghost\nmodule L = FStar.List.Tot\nmodule U32 = FStar.UInt32\nmodule ST = FStar.HyperStack.ST\n\nopen FStar.HyperStack.ST\n\n/// Breaking from my own recommendation here... using fuel 1 / ifuel 1 since\n/// everything does recursive list reasoning.\n#set-options \"--__no_positivity --fuel 1 --ifuel 1\"\n\n/// Definition of the Low* type\n/// ---------------------------\nnoeq\ntype t (a: Type0) =\n B.pointer_or_null (cell a)\n\nand cell (a: Type0) = {\n next: t a;\n data: a;\n}\n\n/// Unlike the canonical style that associates a ``v`` function to any stateful\n/// representation, we use a relation here. Writing a function is just\n/// impossible, since we can't prove termination of ``v`` owing to the fact that\n/// there may be an infinite number of cells in the heap.\n///\n/// LL2 is the module that switches to the function style.\n///\n/// Note: no need to use erased here for ``l`` since the function is already in ``GTot``.\nlet rec well_formed #a (h: HS.mem) (c: t a) (l: list a): GTot Type0 (decreases l) =\n B.live h c /\\ (\n match l with\n | [] ->\n B.g_is_null c\n | a :: q ->\n B.length c == 1 /\\ (\n let { data=data; next=next } = B.get h c 0 in\n a == data /\\\n well_formed h next (G.hide q)\n ))\n\n/// Precise reasoning via the cells predicate\n/// -----------------------------------------\n///\n/// The modifies clauses are given either on the footprint of the list, or on a\n/// region that includes the footprint of the list (higher up the stack, e.g.\n/// LL2). This is good, but does not rule out \"stupid\" implementations that\n/// might re-allocate every single list cell when, say, doing a pop.\n///\n/// For advanced usages which take pointers directly to individual list cells,\n/// it's important to rule out these cases. We thus define the ``cells``\n/// predicate for that purpose.\n\nlet rec cells #a (h: HS.mem) (c: t a) (l: list a): Ghost (list (B.pointer (cell a)))\n (requires well_formed h c l)\n (ensures fun _ -> True)\n (decreases l)\n=\n if B.g_is_null c then\n []\n else\n c :: cells h (B.deref h c).next (List.Tot.tl l)\n\n/// A missing bit of automated reasoning\n/// ------------------------------------\n\n/// This should be absolutely trivial, yet, because there's no pattern on\n/// null_unique, we have to do the case analysis and the decomposition\n/// ourselves. This is quite tedious, as it usually takes a while (at least it\n/// did for me!) to realize that this property does not hold automatically.\n///\n/// I'm setting an SMTPat on this lemma because it's absolutely impossible to\n/// get any work done without it.\nlet same_cells_same_pointer #a (h0 h1: HS.mem) (ll0 ll1: t a) (l0 l1: list a): Lemma\n (requires (\n well_formed h0 ll0 l0 /\\\n well_formed h1 ll1 l1 /\\\n cells h0 ll0 l0 == cells h1 ll1 l1))\n (ensures (\n ll0 == ll1))\n [ SMTPat (cells h0 ll0 l0); SMTPat (cells h1 ll1 l1) ]\n=\n if B.g_is_null ll0 && B.g_is_null ll1 then begin\n B.null_unique ll0;\n B.null_unique ll1\n end else if not (B.g_is_null ll0) && not (B.g_is_null ll1) then begin\n ()\n end else\n false_elim ()\n\n/// Classic stateful reasoning lemmas\n/// ---------------------------------\n///\n/// This essentially follows ``stateful``, the type class of a mutable piece of\n/// state defined using the modifies-theory of Low*v2 and equipped with the\n/// right lemmas to enable usage by clients. See\n/// hacl-star/code/streaming/Hacl.Streaming.Interface.fst for more context.\n\nval footprint: (#a: Type) -> (h: HS.mem) -> (l: t a) -> (n: list a) -> Ghost B.loc\n (requires (well_formed h l n))\n (ensures (fun refs -> True))\n (decreases n)\n\nlet rec footprint #a h l n =\n if B.g_is_null l\n then B.loc_none\n else\n let {next = next} = B.get h l 0 in\n let refs = footprint h next (G.hide (L.tl n)) in\n B.loc_union (B.loc_addr_of_buffer l) refs\n\n/// Departing from LinkedList4 here. I prefer to bolt these into the invariant\n/// rather than requiring on some wizard lemmas from LowStar.Monotonic.Buffer to\n/// deduce these.\nval cells_pairwise_disjoint: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_pairwise_disjoint #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.loc_disjoint (B.loc_addr_of_buffer l) (footprint h next (L.tl n)) /\\\n cells_pairwise_disjoint h next (L.tl n)\n\nval cells_live_freeable: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_live_freeable #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.live h l /\\ B.freeable l /\\\n cells_live_freeable h next (L.tl n)\n\nval invariant: #a:Type -> h:HS.mem -> l: t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\nlet invariant #a h l n =\n cells_live_freeable h l n /\\ cells_pairwise_disjoint h l n\n\n/// Normally this would be automatic and writing a custom lemma for that stuff\n/// would be needed only when writing an fsti, however, as stated earlier,\n/// recursion makes everything more difficult and this absolutely needs to be in\n/// scope otherwise the modifies-clause theory cannot deduce disjointness of\n/// fresh allocations w.r.t. the footprint.\nlet rec invariant_loc_in_footprint (#a: Type) (h: HS.mem) (l: t a) (n: list a): Lemma\n (requires well_formed h l n)\n (ensures B.loc_in (footprint h l n) h)\n (decreases n)\n [ SMTPat (invariant h l n) ]\n=\n if B.g_is_null l then\n ()\n else\n let next = (B.deref h l).next in\n invariant_loc_in_footprint h next (L.tl n)\n\n/// Another absolutely essential lemma, for which interestingly the pattern\n/// \"invariant h1 l n\" does not work, but well_formed does... I guess that's\n/// alright.\nlet rec frame (#a: Type) (l: t a) (n: list a) (r: B.loc) (h0 h1: HS.mem): Lemma\n (requires (\n well_formed h0 l n /\\\n invariant h0 l n /\\\n B.loc_disjoint r (footprint h0 l n) /\\\n B.modifies r h0 h1\n ))\n (ensures (\n well_formed h1 l n /\\\n footprint h1 l n == footprint h0 l n /\\\n cells h1 l n == cells h0 l n /\\\n invariant h1 l n\n ))\n (decreases n)\n [ SMTPat (well_formed h1 l n); SMTPat (B.modifies r h0 h1) ]\n=\n if B.g_is_null l then\n ()\n else\n frame (B.deref h0 l).next (L.tl n) r h0 h1\n\n/// Lemmas for working with linked lists\n/// ------------------------------------\n///\n/// Since it's a recursive data structure, a lot of the automated reasoning from\n/// LowStar.Buffer doesn't work since it requires inductions. Some helpful\n/// properties are therefore proved here.\n\nlet rec length_functional #a (h: HS.mem) (c: t a) (l1 l2: list a):\n Lemma\n (requires (well_formed h c l1 /\\ well_formed h c l2))\n (ensures (l1 == l2))\n (decreases (G.reveal l1))\n [ SMTPat (well_formed h c l1); SMTPat (well_formed h c l2) ] =\n if B.g_is_null c\n then ()\n else\n let { next=next } = B.get h c 0 in\n length_functional h next (G.hide (L.tl l1)) (G.hide (L.tl l2))\n\n/// The footprint is based on loc_addr_of_buffer so that we can write a free\n/// operation that operates on the footprint. However, this invalidates most of\n/// the helper lemmas for disjointness that were present in the previous\n/// iteration of this module, named LinkedList4.fst. It's ok, we bake\n/// disjointness directly onto the invariant and it works just as well.\n\n/// Connection between ``cells`` and ``v``\n/// --------------------------------------\n\n/// Redefining some helpers.\n#push-options \"--ifuel 1\"\nlet rec gmap #a #b (f: a -> GTot b) (xs: list a): GTot (list b) =\n match xs with\n | [] -> []\n | x :: xs -> f x :: gmap f xs\n\n/// \"As we all know\", right folds are the easiest to work with because they\n/// follow the natural structure of recursion (left-folds are evil).\nlet rec gfold_right #a #b (f: b -> a -> GTot b) (xs: list a) (acc: b): Ghost b\n (requires True)\n (ensures fun _ -> True)\n (decreases xs)\n=\n match xs with\n | [] -> acc\n | x :: xs -> f (gfold_right f xs acc) x\n#pop-options\n\n/// Connecting ``v`` and ``cells``.\n///\n/// See ``tests/Wireguard.fst`` for why going through gmap is important and how\n/// to structure your definitions (should you need to reason about ``cells``) in\n/// a way that will make your life easier.\nlet deref_data #a (h: HS.mem) (c: B.pointer (cell a)): GTot a =\n (B.deref h c).data\n\n#push-options \"--fuel 1 --ifuel 1\"\nlet rec deref_cells_is_v #a (h: HS.mem) (ll: t a) (l: list a): Lemma\n (requires\n well_formed h ll l /\\\n invariant h ll l)\n (ensures\n gmap (deref_data h) (cells h ll l) == l)\n (decreases l)\n [ SMTPat (well_formed h ll l) ]\n=\n if B.g_is_null ll then\n ()\n else\n deref_cells_is_v h (B.deref h ll).next (List.Tot.tl l)\n#pop-options\n\n/// Connecting ``footprint`` and ``cells``.\n///\n/// This allows conveniently switching to a low-level representation of the\n/// footprint in case clients want to do some ultra-precise reasoning about\n/// what's happening with the list spine. However, the fact that all\n/// operations from this module specify things in terms of ``cells`` along with\n/// ``same_cells_same_pointer`` should cover most common cases. (This predates\n/// the introduction of ``same_cells_same_pointer``.)\n///\n/// So, no SMTPat here. Note, I also used the same trick as above with\n/// higher-order combinators to isolate the use of the argument ``h`` to a\n/// single sub-term rather than having it passed down the recursive calls.\n#push-options \"--fuel 2 --ifuel 2\"\nlet rec footprint_via_cells_is_footprint #a (h: HS.mem) (ll: t a) (l: list a): Lemma\n (requires well_formed h ll l)\n (ensures\n footprint h ll l == (gfold_right B.loc_union (gmap B.loc_addr_of_buffer (cells h ll l)) B.loc_none))\n (decreases l)\n=\n if B.g_is_null ll then\n ()\n else\n footprint_via_cells_is_footprint h (B.deref h ll).next (List.Tot.tl l)\n#pop-options\n\n\n/// Stateful operations\n/// -------------------\n///\n/// One thing to note: we pass the region `r` explicitly, which is certainly not\n/// very convenient, but that's alright, another layer of abstraction will take\n/// care of adding existentials and extra machinery to make this more pleasant\n/// to use.\n///\n/// All of the operations below are low-level (in the sense that they rely on\n/// the predicate). I expect clients to use exclusively the variants of these\n/// functions present in LL2.\n\nval push: (#a: Type) -> (r: HS.rid) -> (n: G.erased (list a)) -> (pl: B.pointer (t a)) -> (x: a) ->\n ST unit\n (requires (fun h ->\n let l = B.deref h pl in\n B.live h pl /\\\n well_formed h l n /\\\n invariant h l n /\\\n ST.is_eternal_region r /\\\n B.(loc_includes (loc_region_only true r) (footprint h l n)) /\\\n B.(loc_disjoint (loc_buffer pl) (loc_region_only true r))\n ))\n (ensures (fun h0 _ h1 ->\n let n' = G.hide (x :: G.reveal n) in\n let l = B.deref h1 pl in\n // Style note: I don't repeat ``B.live pl`` in the post-condition since\n // ``B.modifies (loc_buffer pl) h0 h1`` implies that ``B.live h1 pl``.\n B.modifies (B.loc_buffer pl) h0 h1 /\\\n well_formed h1 l n' /\\\n invariant h1 l n' /\\\n B.(loc_includes (loc_region_only true r) (footprint h1 l n') /\\\n Cons? (cells h1 l n') /\\ List.Tot.tail (cells h1 l n') == cells h0 (B.deref h0 pl) n /\\\n B.fresh_loc (B.loc_addr_of_buffer (List.Tot.hd (cells h1 l n'))) h0 h1)\n ))\n\nPartial implementation:\nlet push #a r n pl x =\n \n"
},
{
"role": "assistant",
"content": "let h0 = ST.get () in\nlet l = !*pl in\nlet c = { data = x; next = l } in\nlet pc:B.pointer (cell a) = B.malloc r c 1ul in\nlet h1 = ST.get () in\n(let open B in modifies_only_not_unused_in loc_none h0 h1);\nassert B.(loc_disjoint (loc_buffer pc) (footprint h0 l n));\npl *= pc;\nlet h2 = ST.get () in\nlet n' = G.hide (x :: G.reveal n) in\n(let open B in modifies_trans loc_none h0 h1 (loc_buffer pl) h2);\nassert (well_formed h2 (B.deref h2 pl) n');\nassert (invariant h2 (B.deref h2 pl) n');\nassert ((B.deref h2 (B.deref h2 pl)).next == l);\n()"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: _: Prims.unit -> FStar.HyperStack.ST.ST FStar.Int32.t\nEffect: FStar.HyperStack.ST.ST\n\nContext:\nmodule LowStar.Lib.LinkedList\n\n/// This module, oftentimes referred to as LL1 in other parts of the\n/// documentation, provides a low-level view over a linked list. Unless you\n/// intend on modifying the structure of the linked list (e.g. removing cells)\n/// or iterating over each cell, your are probably better off using LL2.\n///\n/// This module is intentionally not relying on a tight abstraction, to i) permit\n/// direct-style iteration by clients and ii) allow stating some SMTPat lemmas\n/// only on LL1 rather than replicating them in LL2 (going from LL2 to LL1 can\n/// happen automatically at the SMT level since LL2 does not have an fsti either).\n///\n/// Some open questions relative to the design of this module:\n/// - should ``v`` be defined directly in terms of ``cells`` rather than have a\n/// recursive definition along with an equivalence lemma?\n/// - should ``footprint`` be defined directly in terms of ``cells`` rather than\n/// have a recursive definition along with an equivalence lemma?\n/// - is it worth the trouble defining a footprint? my intuition is that it's\n/// fine and allows some clients to do very precise reasoning -- other modules\n/// (e.g. AssocList), when establishing abstraction boundaries, can simplify\n/// things and get rid of the precise footprint reasoning in favor of a static\n/// region-based footprint\n///\n/// Some remaining work items:\n/// - switch to C.Loops.while\n\nopen LowStar.BufferOps\n\nmodule B = LowStar.Buffer\nmodule HS = FStar.HyperStack\nmodule G = FStar.Ghost\nmodule L = FStar.List.Tot\nmodule U32 = FStar.UInt32\nmodule ST = FStar.HyperStack.ST\n\nopen FStar.HyperStack.ST\n\n/// Breaking from my own recommendation here... using fuel 1 / ifuel 1 since\n/// everything does recursive list reasoning.\n#set-options \"--__no_positivity --fuel 1 --ifuel 1\"\n\n/// Definition of the Low* type\n/// ---------------------------\nnoeq\ntype t (a: Type0) =\n B.pointer_or_null (cell a)\n\nand cell (a: Type0) = {\n next: t a;\n data: a;\n}\n\n/// Unlike the canonical style that associates a ``v`` function to any stateful\n/// representation, we use a relation here. Writing a function is just\n/// impossible, since we can't prove termination of ``v`` owing to the fact that\n/// there may be an infinite number of cells in the heap.\n///\n/// LL2 is the module that switches to the function style.\n///\n/// Note: no need to use erased here for ``l`` since the function is already in ``GTot``.\nlet rec well_formed #a (h: HS.mem) (c: t a) (l: list a): GTot Type0 (decreases l) =\n B.live h c /\\ (\n match l with\n | [] ->\n B.g_is_null c\n | a :: q ->\n B.length c == 1 /\\ (\n let { data=data; next=next } = B.get h c 0 in\n a == data /\\\n well_formed h next (G.hide q)\n ))\n\n/// Precise reasoning via the cells predicate\n/// -----------------------------------------\n///\n/// The modifies clauses are given either on the footprint of the list, or on a\n/// region that includes the footprint of the list (higher up the stack, e.g.\n/// LL2). This is good, but does not rule out \"stupid\" implementations that\n/// might re-allocate every single list cell when, say, doing a pop.\n///\n/// For advanced usages which take pointers directly to individual list cells,\n/// it's important to rule out these cases. We thus define the ``cells``\n/// predicate for that purpose.\n\nlet rec cells #a (h: HS.mem) (c: t a) (l: list a): Ghost (list (B.pointer (cell a)))\n (requires well_formed h c l)\n (ensures fun _ -> True)\n (decreases l)\n=\n if B.g_is_null c then\n []\n else\n c :: cells h (B.deref h c).next (List.Tot.tl l)\n\n/// A missing bit of automated reasoning\n/// ------------------------------------\n\n/// This should be absolutely trivial, yet, because there's no pattern on\n/// null_unique, we have to do the case analysis and the decomposition\n/// ourselves. This is quite tedious, as it usually takes a while (at least it\n/// did for me!) to realize that this property does not hold automatically.\n///\n/// I'm setting an SMTPat on this lemma because it's absolutely impossible to\n/// get any work done without it.\nlet same_cells_same_pointer #a (h0 h1: HS.mem) (ll0 ll1: t a) (l0 l1: list a): Lemma\n (requires (\n well_formed h0 ll0 l0 /\\\n well_formed h1 ll1 l1 /\\\n cells h0 ll0 l0 == cells h1 ll1 l1))\n (ensures (\n ll0 == ll1))\n [ SMTPat (cells h0 ll0 l0); SMTPat (cells h1 ll1 l1) ]\n=\n if B.g_is_null ll0 && B.g_is_null ll1 then begin\n B.null_unique ll0;\n B.null_unique ll1\n end else if not (B.g_is_null ll0) && not (B.g_is_null ll1) then begin\n ()\n end else\n false_elim ()\n\n/// Classic stateful reasoning lemmas\n/// ---------------------------------\n///\n/// This essentially follows ``stateful``, the type class of a mutable piece of\n/// state defined using the modifies-theory of Low*v2 and equipped with the\n/// right lemmas to enable usage by clients. See\n/// hacl-star/code/streaming/Hacl.Streaming.Interface.fst for more context.\n\nval footprint: (#a: Type) -> (h: HS.mem) -> (l: t a) -> (n: list a) -> Ghost B.loc\n (requires (well_formed h l n))\n (ensures (fun refs -> True))\n (decreases n)\n\nlet rec footprint #a h l n =\n if B.g_is_null l\n then B.loc_none\n else\n let {next = next} = B.get h l 0 in\n let refs = footprint h next (G.hide (L.tl n)) in\n B.loc_union (B.loc_addr_of_buffer l) refs\n\n/// Departing from LinkedList4 here. I prefer to bolt these into the invariant\n/// rather than requiring on some wizard lemmas from LowStar.Monotonic.Buffer to\n/// deduce these.\nval cells_pairwise_disjoint: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_pairwise_disjoint #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.loc_disjoint (B.loc_addr_of_buffer l) (footprint h next (L.tl n)) /\\\n cells_pairwise_disjoint h next (L.tl n)\n\nval cells_live_freeable: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_live_freeable #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.live h l /\\ B.freeable l /\\\n cells_live_freeable h next (L.tl n)\n\nval invariant: #a:Type -> h:HS.mem -> l: t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\nlet invariant #a h l n =\n cells_live_freeable h l n /\\ cells_pairwise_disjoint h l n\n\n/// Normally this would be automatic and writing a custom lemma for that stuff\n/// would be needed only when writing an fsti, however, as stated earlier,\n/// recursion makes everything more difficult and this absolutely needs to be in\n/// scope otherwise the modifies-clause theory cannot deduce disjointness of\n/// fresh allocations w.r.t. the footprint.\nlet rec invariant_loc_in_footprint (#a: Type) (h: HS.mem) (l: t a) (n: list a): Lemma\n (requires well_formed h l n)\n (ensures B.loc_in (footprint h l n) h)\n (decreases n)\n [ SMTPat (invariant h l n) ]\n=\n if B.g_is_null l then\n ()\n else\n let next = (B.deref h l).next in\n invariant_loc_in_footprint h next (L.tl n)\n\n/// Another absolutely essential lemma, for which interestingly the pattern\n/// \"invariant h1 l n\" does not work, but well_formed does... I guess that's\n/// alright.\nlet rec frame (#a: Type) (l: t a) (n: list a) (r: B.loc) (h0 h1: HS.mem): Lemma\n (requires (\n well_formed h0 l n /\\\n invariant h0 l n /\\\n B.loc_disjoint r (footprint h0 l n) /\\\n B.modifies r h0 h1\n ))\n (ensures (\n well_formed h1 l n /\\\n footprint h1 l n == footprint h0 l n /\\\n cells h1 l n == cells h0 l n /\\\n invariant h1 l n\n ))\n (decreases n)\n [ SMTPat (well_formed h1 l n); SMTPat (B.modifies r h0 h1) ]\n=\n if B.g_is_null l then\n ()\n else\n frame (B.deref h0 l).next (L.tl n) r h0 h1\n\n/// Lemmas for working with linked lists\n/// ------------------------------------\n///\n/// Since it's a recursive data structure, a lot of the automated reasoning from\n/// LowStar.Buffer doesn't work since it requires inductions. Some helpful\n/// properties are therefore proved here.\n\nlet rec length_functional #a (h: HS.mem) (c: t a) (l1 l2: list a):\n Lemma\n (requires (well_formed h c l1 /\\ well_formed h c l2))\n (ensures (l1 == l2))\n (decreases (G.reveal l1))\n [ SMTPat (well_formed h c l1); SMTPat (well_formed h c l2) ] =\n if B.g_is_null c\n then ()\n else\n let { next=next } = B.get h c 0 in\n length_functional h next (G.hide (L.tl l1)) (G.hide (L.tl l2))\n\n/// The footprint is based on loc_addr_of_buffer so that we can write a free\n/// operation that operates on the footprint. However, this invalidates most of\n/// the helper lemmas for disjointness that were present in the previous\n/// iteration of this module, named LinkedList4.fst. It's ok, we bake\n/// disjointness directly onto the invariant and it works just as well.\n\n/// Connection between ``cells`` and ``v``\n/// --------------------------------------\n\n/// Redefining some helpers.\n#push-options \"--ifuel 1\"\nlet rec gmap #a #b (f: a -> GTot b) (xs: list a): GTot (list b) =\n match xs with\n | [] -> []\n | x :: xs -> f x :: gmap f xs\n\n/// \"As we all know\", right folds are the easiest to work with because they\n/// follow the natural structure of recursion (left-folds are evil).\nlet rec gfold_right #a #b (f: b -> a -> GTot b) (xs: list a) (acc: b): Ghost b\n (requires True)\n (ensures fun _ -> True)\n (decreases xs)\n=\n match xs with\n | [] -> acc\n | x :: xs -> f (gfold_right f xs acc) x\n#pop-options\n\n/// Connecting ``v`` and ``cells``.\n///\n/// See ``tests/Wireguard.fst`` for why going through gmap is important and how\n/// to structure your definitions (should you need to reason about ``cells``) in\n/// a way that will make your life easier.\nlet deref_data #a (h: HS.mem) (c: B.pointer (cell a)): GTot a =\n (B.deref h c).data\n\n#push-options \"--fuel 1 --ifuel 1\"\nlet rec deref_cells_is_v #a (h: HS.mem) (ll: t a) (l: list a): Lemma\n (requires\n well_formed h ll l /\\\n invariant h ll l)\n (ensures\n gmap (deref_data h) (cells h ll l) == l)\n (decreases l)\n [ SMTPat (well_formed h ll l) ]\n=\n if B.g_is_null ll then\n ()\n else\n deref_cells_is_v h (B.deref h ll).next (List.Tot.tl l)\n#pop-options\n\n/// Connecting ``footprint`` and ``cells``.\n///\n/// This allows conveniently switching to a low-level representation of the\n/// footprint in case clients want to do some ultra-precise reasoning about\n/// what's happening with the list spine. However, the fact that all\n/// operations from this module specify things in terms of ``cells`` along with\n/// ``same_cells_same_pointer`` should cover most common cases. (This predates\n/// the introduction of ``same_cells_same_pointer``.)\n///\n/// So, no SMTPat here. Note, I also used the same trick as above with\n/// higher-order combinators to isolate the use of the argument ``h`` to a\n/// single sub-term rather than having it passed down the recursive calls.\n#push-options \"--fuel 2 --ifuel 2\"\nlet rec footprint_via_cells_is_footprint #a (h: HS.mem) (ll: t a) (l: list a): Lemma\n (requires well_formed h ll l)\n (ensures\n footprint h ll l == (gfold_right B.loc_union (gmap B.loc_addr_of_buffer (cells h ll l)) B.loc_none))\n (decreases l)\n=\n if B.g_is_null ll then\n ()\n else\n footprint_via_cells_is_footprint h (B.deref h ll).next (List.Tot.tl l)\n#pop-options\n\n\n/// Stateful operations\n/// -------------------\n///\n/// One thing to note: we pass the region `r` explicitly, which is certainly not\n/// very convenient, but that's alright, another layer of abstraction will take\n/// care of adding existentials and extra machinery to make this more pleasant\n/// to use.\n///\n/// All of the operations below are low-level (in the sense that they rely on\n/// the predicate). I expect clients to use exclusively the variants of these\n/// functions present in LL2.\n\nval push: (#a: Type) -> (r: HS.rid) -> (n: G.erased (list a)) -> (pl: B.pointer (t a)) -> (x: a) ->\n ST unit\n (requires (fun h ->\n let l = B.deref h pl in\n B.live h pl /\\\n well_formed h l n /\\\n invariant h l n /\\\n ST.is_eternal_region r /\\\n B.(loc_includes (loc_region_only true r) (footprint h l n)) /\\\n B.(loc_disjoint (loc_buffer pl) (loc_region_only true r))\n ))\n (ensures (fun h0 _ h1 ->\n let n' = G.hide (x :: G.reveal n) in\n let l = B.deref h1 pl in\n // Style note: I don't repeat ``B.live pl`` in the post-condition since\n // ``B.modifies (loc_buffer pl) h0 h1`` implies that ``B.live h1 pl``.\n B.modifies (B.loc_buffer pl) h0 h1 /\\\n well_formed h1 l n' /\\\n invariant h1 l n' /\\\n B.(loc_includes (loc_region_only true r) (footprint h1 l n') /\\\n Cons? (cells h1 l n') /\\ List.Tot.tail (cells h1 l n') == cells h0 (B.deref h0 pl) n /\\\n B.fresh_loc (B.loc_addr_of_buffer (List.Tot.hd (cells h1 l n'))) h0 h1)\n ))\n\nlet push #a r n pl x =\n (**) let h0 = ST.get () in\n let l = !* pl in\n let c = { data = x; next = l } in\n\n let pc: B.pointer (cell a) = B.malloc r c 1ul in\n (**) let h1 = ST.get () in\n (**) B.(modifies_only_not_unused_in loc_none h0 h1);\n (**) assert B.(loc_disjoint (loc_buffer pc) (footprint h0 l n));\n\n pl *= pc;\n (**) let h2 = ST.get () in\n (**) let n' = G.hide (x :: G.reveal n) in\n (**) B.(modifies_trans loc_none h0 h1 (loc_buffer pl) h2);\n (**) assert (well_formed h2 (B.deref h2 pl) n');\n (**) assert (invariant h2 (B.deref h2 pl) n');\n (**) assert ((B.deref h2 (B.deref h2 pl)).next == l);\n\n ()\n\nval pop: (#a: Type) -> (r: HS.rid) -> (n: G.erased (list a)) -> (pl: B.pointer (t a)) ->\n ST a\n (requires (fun h ->\n let l = B.deref h pl in\n B.live h pl /\\\n well_formed h l n /\\\n invariant h l n /\\\n ST.is_eternal_region r /\\\n B.(loc_includes (loc_region_only true r) (footprint h l n)) /\\\n B.loc_disjoint (B.loc_buffer pl) (footprint h l n) /\\\n Cons? n\n ))\n (ensures (fun h0 x h1 ->\n let l = B.deref h1 pl in\n let n' = L.tl n in\n x == L.hd n /\\\n // Introducing a super precise modifies clause (e.g. loc_addr_of_buffer\n // (B.deref h0 pl)) here is not useful and prevents trigger-based\n // reasoning, while also requiring clients to unroll footprint. There's a\n // tension here between revealing that ``pop`` does nothing stupid (e.g.\n // re-allocating all the list cells) and providing an abstract enough\n // predicate to work with.\n B.(modifies (loc_buffer pl `loc_union` footprint h0 (B.deref h0 pl) n) h0 h1) /\\\n well_formed h1 l n' /\\\n invariant h1 l n' /\\\n B.(loc_includes (loc_region_only true r) (footprint h1 l n')) /\\\n List.Tot.tail (cells h0 (B.deref h0 pl) n) == cells h1 l n'))\n\nlet pop #a r n pl =\n let l = !* pl in\n let r = (!*l).data in\n let next = (!*l).next in\n (**) let h1 = ST.get () in\n (**) assert (cells h1 l n == l :: cells h1 next (List.Tot.tl n));\n (**) assert B.(loc_disjoint (loc_buffer pl) (footprint h1 l n));\n (**) assert B.(loc_disjoint (loc_buffer pl) (footprint h1 next (List.Tot.tl n)));\n\n pl *= next;\n B.free l;\n r\n\nval free_: (#a: Type) -> (#n: G.erased (list a)) -> (l: t a) ->\n ST unit\n (requires (fun h ->\n well_formed h l n /\\\n invariant h l n\n ))\n (ensures (fun h0 _ h1 ->\n B.(modifies (footprint h0 l n) h0 h1)))\n\nlet rec free_ #a #n l =\n if B.is_null l then\n ()\n else begin\n let tl: G.erased (list a) = G.hide (L.tl (G.reveal n)) in\n free_ #_ #tl (!*l).next;\n B.free l\n end\n\nval free: (#a: Type) -> (#n: G.erased (list a)) -> (pl: B.pointer (t a)) ->\n ST unit\n (requires (fun h ->\n let l = B.deref h pl in\n B.live h pl /\\\n well_formed h l n /\\\n invariant h l n /\\\n B.loc_disjoint (B.loc_buffer pl) (footprint h l n)\n ))\n (ensures (fun h0 _ h1 ->\n let l = B.deref h1 pl in\n well_formed h1 l [] /\\\n invariant h1 l [] /\\\n footprint h1 l [] == B.loc_none /\\\n cells h1 l [] == [] /\\\n B.(modifies (footprint h0 (B.deref h0 pl) n `loc_union` loc_buffer pl) h0 h1)))\n\nlet free #a #n pl =\n free_ #_ #n !*pl;\n pl *= B.null\n\nval length (#a: Type) (gn: G.erased (list a)) (l: t a): Stack UInt32.t\n (requires (fun h -> well_formed h l gn))\n (ensures (fun h0 n h1 ->\n h0 == h1 /\\\n U32.v n = L.length (G.reveal gn)\n ))\n\nlet rec length #a gn l =\n if B.is_null l then\n 0ul\n else\n let open U32 in\n let c = !* l in\n let next = c.next in\n let n = length (G.hide (L.tail (G.reveal gn))) next in\n if n = 0xfffffffful then begin\n LowStar.Failure.failwith \"Integer overflow in LowStar.LinkedList.length\"\n end else\n n +^ 1ul\n\n/// Small test\n/// ----------\n\nval test: unit -> ST (Int32.t) (fun _ -> true) (fun _ _ _ -> true)\n\nPartial implementation:\nlet test () =\n \n"
},
{
"role": "assistant",
"content": "let l:B.pointer_or_null (t Int32.t) = B.malloc HS.root B.null 1ul in\nlet l_region = new_region HS.root in\npush #Int32.t l_region (G.hide []) l 1l;\npush #Int32.t l_region (G.hide [1l]) l 0l;\nlet r = pop #Int32.t l_region (G.hide [0l; 1l]) l in\nTestLib.checku32 (length (G.hide [1l]) !*l) 1ul;\nr"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: pl: LowStar.Buffer.pointer (LowStar.Lib.LinkedList.t a) -> FStar.HyperStack.ST.ST Prims.unit\nEffect: FStar.HyperStack.ST.ST\n\nContext:\nmodule LowStar.Lib.LinkedList\n\n/// This module, oftentimes referred to as LL1 in other parts of the\n/// documentation, provides a low-level view over a linked list. Unless you\n/// intend on modifying the structure of the linked list (e.g. removing cells)\n/// or iterating over each cell, your are probably better off using LL2.\n///\n/// This module is intentionally not relying on a tight abstraction, to i) permit\n/// direct-style iteration by clients and ii) allow stating some SMTPat lemmas\n/// only on LL1 rather than replicating them in LL2 (going from LL2 to LL1 can\n/// happen automatically at the SMT level since LL2 does not have an fsti either).\n///\n/// Some open questions relative to the design of this module:\n/// - should ``v`` be defined directly in terms of ``cells`` rather than have a\n/// recursive definition along with an equivalence lemma?\n/// - should ``footprint`` be defined directly in terms of ``cells`` rather than\n/// have a recursive definition along with an equivalence lemma?\n/// - is it worth the trouble defining a footprint? my intuition is that it's\n/// fine and allows some clients to do very precise reasoning -- other modules\n/// (e.g. AssocList), when establishing abstraction boundaries, can simplify\n/// things and get rid of the precise footprint reasoning in favor of a static\n/// region-based footprint\n///\n/// Some remaining work items:\n/// - switch to C.Loops.while\n\nopen LowStar.BufferOps\n\nmodule B = LowStar.Buffer\nmodule HS = FStar.HyperStack\nmodule G = FStar.Ghost\nmodule L = FStar.List.Tot\nmodule U32 = FStar.UInt32\nmodule ST = FStar.HyperStack.ST\n\nopen FStar.HyperStack.ST\n\n/// Breaking from my own recommendation here... using fuel 1 / ifuel 1 since\n/// everything does recursive list reasoning.\n#set-options \"--__no_positivity --fuel 1 --ifuel 1\"\n\n/// Definition of the Low* type\n/// ---------------------------\nnoeq\ntype t (a: Type0) =\n B.pointer_or_null (cell a)\n\nand cell (a: Type0) = {\n next: t a;\n data: a;\n}\n\n/// Unlike the canonical style that associates a ``v`` function to any stateful\n/// representation, we use a relation here. Writing a function is just\n/// impossible, since we can't prove termination of ``v`` owing to the fact that\n/// there may be an infinite number of cells in the heap.\n///\n/// LL2 is the module that switches to the function style.\n///\n/// Note: no need to use erased here for ``l`` since the function is already in ``GTot``.\nlet rec well_formed #a (h: HS.mem) (c: t a) (l: list a): GTot Type0 (decreases l) =\n B.live h c /\\ (\n match l with\n | [] ->\n B.g_is_null c\n | a :: q ->\n B.length c == 1 /\\ (\n let { data=data; next=next } = B.get h c 0 in\n a == data /\\\n well_formed h next (G.hide q)\n ))\n\n/// Precise reasoning via the cells predicate\n/// -----------------------------------------\n///\n/// The modifies clauses are given either on the footprint of the list, or on a\n/// region that includes the footprint of the list (higher up the stack, e.g.\n/// LL2). This is good, but does not rule out \"stupid\" implementations that\n/// might re-allocate every single list cell when, say, doing a pop.\n///\n/// For advanced usages which take pointers directly to individual list cells,\n/// it's important to rule out these cases. We thus define the ``cells``\n/// predicate for that purpose.\n\nlet rec cells #a (h: HS.mem) (c: t a) (l: list a): Ghost (list (B.pointer (cell a)))\n (requires well_formed h c l)\n (ensures fun _ -> True)\n (decreases l)\n=\n if B.g_is_null c then\n []\n else\n c :: cells h (B.deref h c).next (List.Tot.tl l)\n\n/// A missing bit of automated reasoning\n/// ------------------------------------\n\n/// This should be absolutely trivial, yet, because there's no pattern on\n/// null_unique, we have to do the case analysis and the decomposition\n/// ourselves. This is quite tedious, as it usually takes a while (at least it\n/// did for me!) to realize that this property does not hold automatically.\n///\n/// I'm setting an SMTPat on this lemma because it's absolutely impossible to\n/// get any work done without it.\nlet same_cells_same_pointer #a (h0 h1: HS.mem) (ll0 ll1: t a) (l0 l1: list a): Lemma\n (requires (\n well_formed h0 ll0 l0 /\\\n well_formed h1 ll1 l1 /\\\n cells h0 ll0 l0 == cells h1 ll1 l1))\n (ensures (\n ll0 == ll1))\n [ SMTPat (cells h0 ll0 l0); SMTPat (cells h1 ll1 l1) ]\n=\n if B.g_is_null ll0 && B.g_is_null ll1 then begin\n B.null_unique ll0;\n B.null_unique ll1\n end else if not (B.g_is_null ll0) && not (B.g_is_null ll1) then begin\n ()\n end else\n false_elim ()\n\n/// Classic stateful reasoning lemmas\n/// ---------------------------------\n///\n/// This essentially follows ``stateful``, the type class of a mutable piece of\n/// state defined using the modifies-theory of Low*v2 and equipped with the\n/// right lemmas to enable usage by clients. See\n/// hacl-star/code/streaming/Hacl.Streaming.Interface.fst for more context.\n\nval footprint: (#a: Type) -> (h: HS.mem) -> (l: t a) -> (n: list a) -> Ghost B.loc\n (requires (well_formed h l n))\n (ensures (fun refs -> True))\n (decreases n)\n\nlet rec footprint #a h l n =\n if B.g_is_null l\n then B.loc_none\n else\n let {next = next} = B.get h l 0 in\n let refs = footprint h next (G.hide (L.tl n)) in\n B.loc_union (B.loc_addr_of_buffer l) refs\n\n/// Departing from LinkedList4 here. I prefer to bolt these into the invariant\n/// rather than requiring on some wizard lemmas from LowStar.Monotonic.Buffer to\n/// deduce these.\nval cells_pairwise_disjoint: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_pairwise_disjoint #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.loc_disjoint (B.loc_addr_of_buffer l) (footprint h next (L.tl n)) /\\\n cells_pairwise_disjoint h next (L.tl n)\n\nval cells_live_freeable: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_live_freeable #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.live h l /\\ B.freeable l /\\\n cells_live_freeable h next (L.tl n)\n\nval invariant: #a:Type -> h:HS.mem -> l: t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\nlet invariant #a h l n =\n cells_live_freeable h l n /\\ cells_pairwise_disjoint h l n\n\n/// Normally this would be automatic and writing a custom lemma for that stuff\n/// would be needed only when writing an fsti, however, as stated earlier,\n/// recursion makes everything more difficult and this absolutely needs to be in\n/// scope otherwise the modifies-clause theory cannot deduce disjointness of\n/// fresh allocations w.r.t. the footprint.\nlet rec invariant_loc_in_footprint (#a: Type) (h: HS.mem) (l: t a) (n: list a): Lemma\n (requires well_formed h l n)\n (ensures B.loc_in (footprint h l n) h)\n (decreases n)\n [ SMTPat (invariant h l n) ]\n=\n if B.g_is_null l then\n ()\n else\n let next = (B.deref h l).next in\n invariant_loc_in_footprint h next (L.tl n)\n\n/// Another absolutely essential lemma, for which interestingly the pattern\n/// \"invariant h1 l n\" does not work, but well_formed does... I guess that's\n/// alright.\nlet rec frame (#a: Type) (l: t a) (n: list a) (r: B.loc) (h0 h1: HS.mem): Lemma\n (requires (\n well_formed h0 l n /\\\n invariant h0 l n /\\\n B.loc_disjoint r (footprint h0 l n) /\\\n B.modifies r h0 h1\n ))\n (ensures (\n well_formed h1 l n /\\\n footprint h1 l n == footprint h0 l n /\\\n cells h1 l n == cells h0 l n /\\\n invariant h1 l n\n ))\n (decreases n)\n [ SMTPat (well_formed h1 l n); SMTPat (B.modifies r h0 h1) ]\n=\n if B.g_is_null l then\n ()\n else\n frame (B.deref h0 l).next (L.tl n) r h0 h1\n\n/// Lemmas for working with linked lists\n/// ------------------------------------\n///\n/// Since it's a recursive data structure, a lot of the automated reasoning from\n/// LowStar.Buffer doesn't work since it requires inductions. Some helpful\n/// properties are therefore proved here.\n\nlet rec length_functional #a (h: HS.mem) (c: t a) (l1 l2: list a):\n Lemma\n (requires (well_formed h c l1 /\\ well_formed h c l2))\n (ensures (l1 == l2))\n (decreases (G.reveal l1))\n [ SMTPat (well_formed h c l1); SMTPat (well_formed h c l2) ] =\n if B.g_is_null c\n then ()\n else\n let { next=next } = B.get h c 0 in\n length_functional h next (G.hide (L.tl l1)) (G.hide (L.tl l2))\n\n/// The footprint is based on loc_addr_of_buffer so that we can write a free\n/// operation that operates on the footprint. However, this invalidates most of\n/// the helper lemmas for disjointness that were present in the previous\n/// iteration of this module, named LinkedList4.fst. It's ok, we bake\n/// disjointness directly onto the invariant and it works just as well.\n\n/// Connection between ``cells`` and ``v``\n/// --------------------------------------\n\n/// Redefining some helpers.\n#push-options \"--ifuel 1\"\nlet rec gmap #a #b (f: a -> GTot b) (xs: list a): GTot (list b) =\n match xs with\n | [] -> []\n | x :: xs -> f x :: gmap f xs\n\n/// \"As we all know\", right folds are the easiest to work with because they\n/// follow the natural structure of recursion (left-folds are evil).\nlet rec gfold_right #a #b (f: b -> a -> GTot b) (xs: list a) (acc: b): Ghost b\n (requires True)\n (ensures fun _ -> True)\n (decreases xs)\n=\n match xs with\n | [] -> acc\n | x :: xs -> f (gfold_right f xs acc) x\n#pop-options\n\n/// Connecting ``v`` and ``cells``.\n///\n/// See ``tests/Wireguard.fst`` for why going through gmap is important and how\n/// to structure your definitions (should you need to reason about ``cells``) in\n/// a way that will make your life easier.\nlet deref_data #a (h: HS.mem) (c: B.pointer (cell a)): GTot a =\n (B.deref h c).data\n\n#push-options \"--fuel 1 --ifuel 1\"\nlet rec deref_cells_is_v #a (h: HS.mem) (ll: t a) (l: list a): Lemma\n (requires\n well_formed h ll l /\\\n invariant h ll l)\n (ensures\n gmap (deref_data h) (cells h ll l) == l)\n (decreases l)\n [ SMTPat (well_formed h ll l) ]\n=\n if B.g_is_null ll then\n ()\n else\n deref_cells_is_v h (B.deref h ll).next (List.Tot.tl l)\n#pop-options\n\n/// Connecting ``footprint`` and ``cells``.\n///\n/// This allows conveniently switching to a low-level representation of the\n/// footprint in case clients want to do some ultra-precise reasoning about\n/// what's happening with the list spine. However, the fact that all\n/// operations from this module specify things in terms of ``cells`` along with\n/// ``same_cells_same_pointer`` should cover most common cases. (This predates\n/// the introduction of ``same_cells_same_pointer``.)\n///\n/// So, no SMTPat here. Note, I also used the same trick as above with\n/// higher-order combinators to isolate the use of the argument ``h`` to a\n/// single sub-term rather than having it passed down the recursive calls.\n#push-options \"--fuel 2 --ifuel 2\"\nlet rec footprint_via_cells_is_footprint #a (h: HS.mem) (ll: t a) (l: list a): Lemma\n (requires well_formed h ll l)\n (ensures\n footprint h ll l == (gfold_right B.loc_union (gmap B.loc_addr_of_buffer (cells h ll l)) B.loc_none))\n (decreases l)\n=\n if B.g_is_null ll then\n ()\n else\n footprint_via_cells_is_footprint h (B.deref h ll).next (List.Tot.tl l)\n#pop-options\n\n\n/// Stateful operations\n/// -------------------\n///\n/// One thing to note: we pass the region `r` explicitly, which is certainly not\n/// very convenient, but that's alright, another layer of abstraction will take\n/// care of adding existentials and extra machinery to make this more pleasant\n/// to use.\n///\n/// All of the operations below are low-level (in the sense that they rely on\n/// the predicate). I expect clients to use exclusively the variants of these\n/// functions present in LL2.\n\nval push: (#a: Type) -> (r: HS.rid) -> (n: G.erased (list a)) -> (pl: B.pointer (t a)) -> (x: a) ->\n ST unit\n (requires (fun h ->\n let l = B.deref h pl in\n B.live h pl /\\\n well_formed h l n /\\\n invariant h l n /\\\n ST.is_eternal_region r /\\\n B.(loc_includes (loc_region_only true r) (footprint h l n)) /\\\n B.(loc_disjoint (loc_buffer pl) (loc_region_only true r))\n ))\n (ensures (fun h0 _ h1 ->\n let n' = G.hide (x :: G.reveal n) in\n let l = B.deref h1 pl in\n // Style note: I don't repeat ``B.live pl`` in the post-condition since\n // ``B.modifies (loc_buffer pl) h0 h1`` implies that ``B.live h1 pl``.\n B.modifies (B.loc_buffer pl) h0 h1 /\\\n well_formed h1 l n' /\\\n invariant h1 l n' /\\\n B.(loc_includes (loc_region_only true r) (footprint h1 l n') /\\\n Cons? (cells h1 l n') /\\ List.Tot.tail (cells h1 l n') == cells h0 (B.deref h0 pl) n /\\\n B.fresh_loc (B.loc_addr_of_buffer (List.Tot.hd (cells h1 l n'))) h0 h1)\n ))\n\nlet push #a r n pl x =\n (**) let h0 = ST.get () in\n let l = !* pl in\n let c = { data = x; next = l } in\n\n let pc: B.pointer (cell a) = B.malloc r c 1ul in\n (**) let h1 = ST.get () in\n (**) B.(modifies_only_not_unused_in loc_none h0 h1);\n (**) assert B.(loc_disjoint (loc_buffer pc) (footprint h0 l n));\n\n pl *= pc;\n (**) let h2 = ST.get () in\n (**) let n' = G.hide (x :: G.reveal n) in\n (**) B.(modifies_trans loc_none h0 h1 (loc_buffer pl) h2);\n (**) assert (well_formed h2 (B.deref h2 pl) n');\n (**) assert (invariant h2 (B.deref h2 pl) n');\n (**) assert ((B.deref h2 (B.deref h2 pl)).next == l);\n\n ()\n\nval pop: (#a: Type) -> (r: HS.rid) -> (n: G.erased (list a)) -> (pl: B.pointer (t a)) ->\n ST a\n (requires (fun h ->\n let l = B.deref h pl in\n B.live h pl /\\\n well_formed h l n /\\\n invariant h l n /\\\n ST.is_eternal_region r /\\\n B.(loc_includes (loc_region_only true r) (footprint h l n)) /\\\n B.loc_disjoint (B.loc_buffer pl) (footprint h l n) /\\\n Cons? n\n ))\n (ensures (fun h0 x h1 ->\n let l = B.deref h1 pl in\n let n' = L.tl n in\n x == L.hd n /\\\n // Introducing a super precise modifies clause (e.g. loc_addr_of_buffer\n // (B.deref h0 pl)) here is not useful and prevents trigger-based\n // reasoning, while also requiring clients to unroll footprint. There's a\n // tension here between revealing that ``pop`` does nothing stupid (e.g.\n // re-allocating all the list cells) and providing an abstract enough\n // predicate to work with.\n B.(modifies (loc_buffer pl `loc_union` footprint h0 (B.deref h0 pl) n) h0 h1) /\\\n well_formed h1 l n' /\\\n invariant h1 l n' /\\\n B.(loc_includes (loc_region_only true r) (footprint h1 l n')) /\\\n List.Tot.tail (cells h0 (B.deref h0 pl) n) == cells h1 l n'))\n\nlet pop #a r n pl =\n let l = !* pl in\n let r = (!*l).data in\n let next = (!*l).next in\n (**) let h1 = ST.get () in\n (**) assert (cells h1 l n == l :: cells h1 next (List.Tot.tl n));\n (**) assert B.(loc_disjoint (loc_buffer pl) (footprint h1 l n));\n (**) assert B.(loc_disjoint (loc_buffer pl) (footprint h1 next (List.Tot.tl n)));\n\n pl *= next;\n B.free l;\n r\n\nval free_: (#a: Type) -> (#n: G.erased (list a)) -> (l: t a) ->\n ST unit\n (requires (fun h ->\n well_formed h l n /\\\n invariant h l n\n ))\n (ensures (fun h0 _ h1 ->\n B.(modifies (footprint h0 l n) h0 h1)))\n\nlet rec free_ #a #n l =\n if B.is_null l then\n ()\n else begin\n let tl: G.erased (list a) = G.hide (L.tl (G.reveal n)) in\n free_ #_ #tl (!*l).next;\n B.free l\n end\n\nval free: (#a: Type) -> (#n: G.erased (list a)) -> (pl: B.pointer (t a)) ->\n ST unit\n (requires (fun h ->\n let l = B.deref h pl in\n B.live h pl /\\\n well_formed h l n /\\\n invariant h l n /\\\n B.loc_disjoint (B.loc_buffer pl) (footprint h l n)\n ))\n (ensures (fun h0 _ h1 ->\n let l = B.deref h1 pl in\n well_formed h1 l [] /\\\n invariant h1 l [] /\\\n footprint h1 l [] == B.loc_none /\\\n cells h1 l [] == [] /\\\n B.(modifies (footprint h0 (B.deref h0 pl) n `loc_union` loc_buffer pl) h0 h1)))\n\nPartial implementation:\nlet free #a #n pl =\n \n"
},
{
"role": "assistant",
"content": "free_ #_ #n !*pl;\npl *= B.null"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n s: LowParse.Spec.Base.serializer p ->\n p32: LowParse.SLow.Base.parser32 p ->\n sz: Prims.nat ->\n sz32: FStar.UInt32.t{FStar.UInt32.v sz32 == sz}\n -> LowParse.SLow.Base.parser32 (LowParse.Spec.FLData.parse_fldata_strong s sz)\nEffect: Prims.Tot\n\nContext:\nmodule LowParse.SLow.FLData\ninclude LowParse.Spec.FLData\ninclude LowParse.SLow.Combinators\n\nmodule B32 = LowParse.Bytes32\nmodule U32 = FStar.UInt32\n\ninline_for_extraction\nlet parse32_fldata\n (#k: parser_kind)\n (#t: Type)\n (#p: parser k t)\n (p32: parser32 p)\n (sz: nat)\n (sz32: U32.t { U32.v sz32 == sz } )\n: Tot (parser32 (parse_fldata p sz))\n= (fun (input: bytes32) -> ((\n if U32.lt (B32.len input) sz32\n then\n None\n else\n match p32 (B32.b32slice input 0ul sz32) with\n | Some (v, consumed) ->\n\tif consumed = sz32\n\tthen begin\n\t Some (v, consumed)\n\tend else None\n | None -> None\n ) <: (res: option (t * U32.t) { parser32_correct (parse_fldata p sz) input res } )))\n\nPartial implementation:\nlet parse32_fldata_strong\n (#k: parser_kind)\n (#t: Type)\n (#p: parser k t)\n (s: serializer p)\n (p32: parser32 p)\n (sz: nat)\n (sz32: U32.t{U32.v sz32 == sz})\n : Tot (parser32 (parse_fldata_strong s sz)) =\n \n"
},
{
"role": "assistant",
"content": "(fun (input: bytes32) ->\n ((match parse32_fldata p32 sz sz32 input with\n | Some (v, consumed) ->\n assert (parse_fldata_strong_correct s sz (B32.reveal input) (U32.v consumed) v;\n Seq.length (s v) == sz);\n Some ((v <: parse_fldata_strong_t s sz), consumed)\n | None -> None)\n <:\n (res:\n option (parse_fldata_strong_t s sz * U32.t)\n {parser32_correct (parse_fldata_strong s sz) input res})))"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: s32: LowParse.SLow.Base.partial_serializer32 s -> sz: Prims.nat{sz < 4294967296}\n -> LowParse.SLow.Base.serializer32 (LowParse.Spec.FLData.serialize_fldata_strong s sz)\nEffect: Prims.Tot\n\nContext:\nmodule LowParse.SLow.FLData\ninclude LowParse.Spec.FLData\ninclude LowParse.SLow.Combinators\n\nmodule B32 = LowParse.Bytes32\nmodule U32 = FStar.UInt32\n\ninline_for_extraction\nlet parse32_fldata\n (#k: parser_kind)\n (#t: Type)\n (#p: parser k t)\n (p32: parser32 p)\n (sz: nat)\n (sz32: U32.t { U32.v sz32 == sz } )\n: Tot (parser32 (parse_fldata p sz))\n= (fun (input: bytes32) -> ((\n if U32.lt (B32.len input) sz32\n then\n None\n else\n match p32 (B32.b32slice input 0ul sz32) with\n | Some (v, consumed) ->\n\tif consumed = sz32\n\tthen begin\n\t Some (v, consumed)\n\tend else None\n | None -> None\n ) <: (res: option (t * U32.t) { parser32_correct (parse_fldata p sz) input res } )))\n\ninline_for_extraction\nlet parse32_fldata_strong\n (#k: parser_kind)\n (#t: Type)\n (#p: parser k t)\n (s: serializer p)\n (p32: parser32 p)\n (sz: nat)\n (sz32: U32.t { U32.v sz32 == sz } )\n: Tot (parser32 (parse_fldata_strong s sz))\n= (fun (input: bytes32) -> ((\n match parse32_fldata p32 sz sz32 input with\n | Some (v, consumed) ->\n assert (\n parse_fldata_strong_correct s sz (B32.reveal input) (U32.v consumed) v;\n Seq.length (s v) == sz\n );\n Some ((v <: parse_fldata_strong_t s sz), consumed)\n | None -> None\n ) \n <: (res: option (parse_fldata_strong_t s sz * U32.t) { parser32_correct (parse_fldata_strong s sz) input res } )))\n\nPartial implementation:\nlet serialize32_fldata_strong\n (#k: parser_kind)\n (#t: Type)\n (#p: parser k t)\n (#s: serializer p)\n (s32: partial_serializer32 s)\n (sz: nat{sz < 4294967296})\n : Tot (serializer32 (serialize_fldata_strong s sz)) =\n \n"
},
{
"role": "assistant",
"content": "(fun (input: parse_fldata_strong_t s sz) -> s32 input)"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n r: FStar.Monotonic.HyperHeap.rid ->\n n: FStar.Ghost.erased (Prims.list a) ->\n pl: LowStar.Buffer.pointer (LowStar.Lib.LinkedList.t a)\n -> FStar.HyperStack.ST.ST a\nEffect: FStar.HyperStack.ST.ST\n\nContext:\nmodule LowStar.Lib.LinkedList\n\n/// This module, oftentimes referred to as LL1 in other parts of the\n/// documentation, provides a low-level view over a linked list. Unless you\n/// intend on modifying the structure of the linked list (e.g. removing cells)\n/// or iterating over each cell, your are probably better off using LL2.\n///\n/// This module is intentionally not relying on a tight abstraction, to i) permit\n/// direct-style iteration by clients and ii) allow stating some SMTPat lemmas\n/// only on LL1 rather than replicating them in LL2 (going from LL2 to LL1 can\n/// happen automatically at the SMT level since LL2 does not have an fsti either).\n///\n/// Some open questions relative to the design of this module:\n/// - should ``v`` be defined directly in terms of ``cells`` rather than have a\n/// recursive definition along with an equivalence lemma?\n/// - should ``footprint`` be defined directly in terms of ``cells`` rather than\n/// have a recursive definition along with an equivalence lemma?\n/// - is it worth the trouble defining a footprint? my intuition is that it's\n/// fine and allows some clients to do very precise reasoning -- other modules\n/// (e.g. AssocList), when establishing abstraction boundaries, can simplify\n/// things and get rid of the precise footprint reasoning in favor of a static\n/// region-based footprint\n///\n/// Some remaining work items:\n/// - switch to C.Loops.while\n\nopen LowStar.BufferOps\n\nmodule B = LowStar.Buffer\nmodule HS = FStar.HyperStack\nmodule G = FStar.Ghost\nmodule L = FStar.List.Tot\nmodule U32 = FStar.UInt32\nmodule ST = FStar.HyperStack.ST\n\nopen FStar.HyperStack.ST\n\n/// Breaking from my own recommendation here... using fuel 1 / ifuel 1 since\n/// everything does recursive list reasoning.\n#set-options \"--__no_positivity --fuel 1 --ifuel 1\"\n\n/// Definition of the Low* type\n/// ---------------------------\nnoeq\ntype t (a: Type0) =\n B.pointer_or_null (cell a)\n\nand cell (a: Type0) = {\n next: t a;\n data: a;\n}\n\n/// Unlike the canonical style that associates a ``v`` function to any stateful\n/// representation, we use a relation here. Writing a function is just\n/// impossible, since we can't prove termination of ``v`` owing to the fact that\n/// there may be an infinite number of cells in the heap.\n///\n/// LL2 is the module that switches to the function style.\n///\n/// Note: no need to use erased here for ``l`` since the function is already in ``GTot``.\nlet rec well_formed #a (h: HS.mem) (c: t a) (l: list a): GTot Type0 (decreases l) =\n B.live h c /\\ (\n match l with\n | [] ->\n B.g_is_null c\n | a :: q ->\n B.length c == 1 /\\ (\n let { data=data; next=next } = B.get h c 0 in\n a == data /\\\n well_formed h next (G.hide q)\n ))\n\n/// Precise reasoning via the cells predicate\n/// -----------------------------------------\n///\n/// The modifies clauses are given either on the footprint of the list, or on a\n/// region that includes the footprint of the list (higher up the stack, e.g.\n/// LL2). This is good, but does not rule out \"stupid\" implementations that\n/// might re-allocate every single list cell when, say, doing a pop.\n///\n/// For advanced usages which take pointers directly to individual list cells,\n/// it's important to rule out these cases. We thus define the ``cells``\n/// predicate for that purpose.\n\nlet rec cells #a (h: HS.mem) (c: t a) (l: list a): Ghost (list (B.pointer (cell a)))\n (requires well_formed h c l)\n (ensures fun _ -> True)\n (decreases l)\n=\n if B.g_is_null c then\n []\n else\n c :: cells h (B.deref h c).next (List.Tot.tl l)\n\n/// A missing bit of automated reasoning\n/// ------------------------------------\n\n/// This should be absolutely trivial, yet, because there's no pattern on\n/// null_unique, we have to do the case analysis and the decomposition\n/// ourselves. This is quite tedious, as it usually takes a while (at least it\n/// did for me!) to realize that this property does not hold automatically.\n///\n/// I'm setting an SMTPat on this lemma because it's absolutely impossible to\n/// get any work done without it.\nlet same_cells_same_pointer #a (h0 h1: HS.mem) (ll0 ll1: t a) (l0 l1: list a): Lemma\n (requires (\n well_formed h0 ll0 l0 /\\\n well_formed h1 ll1 l1 /\\\n cells h0 ll0 l0 == cells h1 ll1 l1))\n (ensures (\n ll0 == ll1))\n [ SMTPat (cells h0 ll0 l0); SMTPat (cells h1 ll1 l1) ]\n=\n if B.g_is_null ll0 && B.g_is_null ll1 then begin\n B.null_unique ll0;\n B.null_unique ll1\n end else if not (B.g_is_null ll0) && not (B.g_is_null ll1) then begin\n ()\n end else\n false_elim ()\n\n/// Classic stateful reasoning lemmas\n/// ---------------------------------\n///\n/// This essentially follows ``stateful``, the type class of a mutable piece of\n/// state defined using the modifies-theory of Low*v2 and equipped with the\n/// right lemmas to enable usage by clients. See\n/// hacl-star/code/streaming/Hacl.Streaming.Interface.fst for more context.\n\nval footprint: (#a: Type) -> (h: HS.mem) -> (l: t a) -> (n: list a) -> Ghost B.loc\n (requires (well_formed h l n))\n (ensures (fun refs -> True))\n (decreases n)\n\nlet rec footprint #a h l n =\n if B.g_is_null l\n then B.loc_none\n else\n let {next = next} = B.get h l 0 in\n let refs = footprint h next (G.hide (L.tl n)) in\n B.loc_union (B.loc_addr_of_buffer l) refs\n\n/// Departing from LinkedList4 here. I prefer to bolt these into the invariant\n/// rather than requiring on some wizard lemmas from LowStar.Monotonic.Buffer to\n/// deduce these.\nval cells_pairwise_disjoint: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_pairwise_disjoint #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.loc_disjoint (B.loc_addr_of_buffer l) (footprint h next (L.tl n)) /\\\n cells_pairwise_disjoint h next (L.tl n)\n\nval cells_live_freeable: #a:Type -> h:HS.mem -> l:t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\n\nlet rec cells_live_freeable #_ h l n =\n if B.g_is_null l then\n True\n else\n let next = (B.deref h l).next in\n B.live h l /\\ B.freeable l /\\\n cells_live_freeable h next (L.tl n)\n\nval invariant: #a:Type -> h:HS.mem -> l: t a -> n:list a -> Ghost Type0\n (requires well_formed h l n)\n (ensures fun _ -> True)\n (decreases n)\nlet invariant #a h l n =\n cells_live_freeable h l n /\\ cells_pairwise_disjoint h l n\n\n/// Normally this would be automatic and writing a custom lemma for that stuff\n/// would be needed only when writing an fsti, however, as stated earlier,\n/// recursion makes everything more difficult and this absolutely needs to be in\n/// scope otherwise the modifies-clause theory cannot deduce disjointness of\n/// fresh allocations w.r.t. the footprint.\nlet rec invariant_loc_in_footprint (#a: Type) (h: HS.mem) (l: t a) (n: list a): Lemma\n (requires well_formed h l n)\n (ensures B.loc_in (footprint h l n) h)\n (decreases n)\n [ SMTPat (invariant h l n) ]\n=\n if B.g_is_null l then\n ()\n else\n let next = (B.deref h l).next in\n invariant_loc_in_footprint h next (L.tl n)\n\n/// Another absolutely essential lemma, for which interestingly the pattern\n/// \"invariant h1 l n\" does not work, but well_formed does... I guess that's\n/// alright.\nlet rec frame (#a: Type) (l: t a) (n: list a) (r: B.loc) (h0 h1: HS.mem): Lemma\n (requires (\n well_formed h0 l n /\\\n invariant h0 l n /\\\n B.loc_disjoint r (footprint h0 l n) /\\\n B.modifies r h0 h1\n ))\n (ensures (\n well_formed h1 l n /\\\n footprint h1 l n == footprint h0 l n /\\\n cells h1 l n == cells h0 l n /\\\n invariant h1 l n\n ))\n (decreases n)\n [ SMTPat (well_formed h1 l n); SMTPat (B.modifies r h0 h1) ]\n=\n if B.g_is_null l then\n ()\n else\n frame (B.deref h0 l).next (L.tl n) r h0 h1\n\n/// Lemmas for working with linked lists\n/// ------------------------------------\n///\n/// Since it's a recursive data structure, a lot of the automated reasoning from\n/// LowStar.Buffer doesn't work since it requires inductions. Some helpful\n/// properties are therefore proved here.\n\nlet rec length_functional #a (h: HS.mem) (c: t a) (l1 l2: list a):\n Lemma\n (requires (well_formed h c l1 /\\ well_formed h c l2))\n (ensures (l1 == l2))\n (decreases (G.reveal l1))\n [ SMTPat (well_formed h c l1); SMTPat (well_formed h c l2) ] =\n if B.g_is_null c\n then ()\n else\n let { next=next } = B.get h c 0 in\n length_functional h next (G.hide (L.tl l1)) (G.hide (L.tl l2))\n\n/// The footprint is based on loc_addr_of_buffer so that we can write a free\n/// operation that operates on the footprint. However, this invalidates most of\n/// the helper lemmas for disjointness that were present in the previous\n/// iteration of this module, named LinkedList4.fst. It's ok, we bake\n/// disjointness directly onto the invariant and it works just as well.\n\n/// Connection between ``cells`` and ``v``\n/// --------------------------------------\n\n/// Redefining some helpers.\n#push-options \"--ifuel 1\"\nlet rec gmap #a #b (f: a -> GTot b) (xs: list a): GTot (list b) =\n match xs with\n | [] -> []\n | x :: xs -> f x :: gmap f xs\n\n/// \"As we all know\", right folds are the easiest to work with because they\n/// follow the natural structure of recursion (left-folds are evil).\nlet rec gfold_right #a #b (f: b -> a -> GTot b) (xs: list a) (acc: b): Ghost b\n (requires True)\n (ensures fun _ -> True)\n (decreases xs)\n=\n match xs with\n | [] -> acc\n | x :: xs -> f (gfold_right f xs acc) x\n#pop-options\n\n/// Connecting ``v`` and ``cells``.\n///\n/// See ``tests/Wireguard.fst`` for why going through gmap is important and how\n/// to structure your definitions (should you need to reason about ``cells``) in\n/// a way that will make your life easier.\nlet deref_data #a (h: HS.mem) (c: B.pointer (cell a)): GTot a =\n (B.deref h c).data\n\n#push-options \"--fuel 1 --ifuel 1\"\nlet rec deref_cells_is_v #a (h: HS.mem) (ll: t a) (l: list a): Lemma\n (requires\n well_formed h ll l /\\\n invariant h ll l)\n (ensures\n gmap (deref_data h) (cells h ll l) == l)\n (decreases l)\n [ SMTPat (well_formed h ll l) ]\n=\n if B.g_is_null ll then\n ()\n else\n deref_cells_is_v h (B.deref h ll).next (List.Tot.tl l)\n#pop-options\n\n/// Connecting ``footprint`` and ``cells``.\n///\n/// This allows conveniently switching to a low-level representation of the\n/// footprint in case clients want to do some ultra-precise reasoning about\n/// what's happening with the list spine. However, the fact that all\n/// operations from this module specify things in terms of ``cells`` along with\n/// ``same_cells_same_pointer`` should cover most common cases. (This predates\n/// the introduction of ``same_cells_same_pointer``.)\n///\n/// So, no SMTPat here. Note, I also used the same trick as above with\n/// higher-order combinators to isolate the use of the argument ``h`` to a\n/// single sub-term rather than having it passed down the recursive calls.\n#push-options \"--fuel 2 --ifuel 2\"\nlet rec footprint_via_cells_is_footprint #a (h: HS.mem) (ll: t a) (l: list a): Lemma\n (requires well_formed h ll l)\n (ensures\n footprint h ll l == (gfold_right B.loc_union (gmap B.loc_addr_of_buffer (cells h ll l)) B.loc_none))\n (decreases l)\n=\n if B.g_is_null ll then\n ()\n else\n footprint_via_cells_is_footprint h (B.deref h ll).next (List.Tot.tl l)\n#pop-options\n\n\n/// Stateful operations\n/// -------------------\n///\n/// One thing to note: we pass the region `r` explicitly, which is certainly not\n/// very convenient, but that's alright, another layer of abstraction will take\n/// care of adding existentials and extra machinery to make this more pleasant\n/// to use.\n///\n/// All of the operations below are low-level (in the sense that they rely on\n/// the predicate). I expect clients to use exclusively the variants of these\n/// functions present in LL2.\n\nval push: (#a: Type) -> (r: HS.rid) -> (n: G.erased (list a)) -> (pl: B.pointer (t a)) -> (x: a) ->\n ST unit\n (requires (fun h ->\n let l = B.deref h pl in\n B.live h pl /\\\n well_formed h l n /\\\n invariant h l n /\\\n ST.is_eternal_region r /\\\n B.(loc_includes (loc_region_only true r) (footprint h l n)) /\\\n B.(loc_disjoint (loc_buffer pl) (loc_region_only true r))\n ))\n (ensures (fun h0 _ h1 ->\n let n' = G.hide (x :: G.reveal n) in\n let l = B.deref h1 pl in\n // Style note: I don't repeat ``B.live pl`` in the post-condition since\n // ``B.modifies (loc_buffer pl) h0 h1`` implies that ``B.live h1 pl``.\n B.modifies (B.loc_buffer pl) h0 h1 /\\\n well_formed h1 l n' /\\\n invariant h1 l n' /\\\n B.(loc_includes (loc_region_only true r) (footprint h1 l n') /\\\n Cons? (cells h1 l n') /\\ List.Tot.tail (cells h1 l n') == cells h0 (B.deref h0 pl) n /\\\n B.fresh_loc (B.loc_addr_of_buffer (List.Tot.hd (cells h1 l n'))) h0 h1)\n ))\n\nlet push #a r n pl x =\n (**) let h0 = ST.get () in\n let l = !* pl in\n let c = { data = x; next = l } in\n\n let pc: B.pointer (cell a) = B.malloc r c 1ul in\n (**) let h1 = ST.get () in\n (**) B.(modifies_only_not_unused_in loc_none h0 h1);\n (**) assert B.(loc_disjoint (loc_buffer pc) (footprint h0 l n));\n\n pl *= pc;\n (**) let h2 = ST.get () in\n (**) let n' = G.hide (x :: G.reveal n) in\n (**) B.(modifies_trans loc_none h0 h1 (loc_buffer pl) h2);\n (**) assert (well_formed h2 (B.deref h2 pl) n');\n (**) assert (invariant h2 (B.deref h2 pl) n');\n (**) assert ((B.deref h2 (B.deref h2 pl)).next == l);\n\n ()\n\nval pop: (#a: Type) -> (r: HS.rid) -> (n: G.erased (list a)) -> (pl: B.pointer (t a)) ->\n ST a\n (requires (fun h ->\n let l = B.deref h pl in\n B.live h pl /\\\n well_formed h l n /\\\n invariant h l n /\\\n ST.is_eternal_region r /\\\n B.(loc_includes (loc_region_only true r) (footprint h l n)) /\\\n B.loc_disjoint (B.loc_buffer pl) (footprint h l n) /\\\n Cons? n\n ))\n (ensures (fun h0 x h1 ->\n let l = B.deref h1 pl in\n let n' = L.tl n in\n x == L.hd n /\\\n // Introducing a super precise modifies clause (e.g. loc_addr_of_buffer\n // (B.deref h0 pl)) here is not useful and prevents trigger-based\n // reasoning, while also requiring clients to unroll footprint. There's a\n // tension here between revealing that ``pop`` does nothing stupid (e.g.\n // re-allocating all the list cells) and providing an abstract enough\n // predicate to work with.\n B.(modifies (loc_buffer pl `loc_union` footprint h0 (B.deref h0 pl) n) h0 h1) /\\\n well_formed h1 l n' /\\\n invariant h1 l n' /\\\n B.(loc_includes (loc_region_only true r) (footprint h1 l n')) /\\\n List.Tot.tail (cells h0 (B.deref h0 pl) n) == cells h1 l n'))\n\nPartial implementation:\nlet pop #a r n pl =\n \n"
},
{
"role": "assistant",
"content": "let l = !*pl in\nlet r = (!*l).data in\nlet next = (!*l).next in\nlet h1 = ST.get () in\nassert (cells h1 l n == l :: cells h1 next (List.Tot.tl n));\nassert B.(loc_disjoint (loc_buffer pl) (footprint h1 l n));\nassert B.(loc_disjoint (loc_buffer pl) (footprint h1 next (List.Tot.tl n)));\npl *= next;\nB.free l;\nr"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n p32: LowParse.SLow.Base.parser32 p ->\n sz: Prims.nat ->\n sz32: FStar.UInt32.t{FStar.UInt32.v sz32 == sz}\n -> LowParse.SLow.Base.parser32 (LowParse.Spec.FLData.parse_fldata p sz)\nEffect: Prims.Tot\n\nContext:\nmodule LowParse.SLow.FLData\ninclude LowParse.Spec.FLData\ninclude LowParse.SLow.Combinators\n\nmodule B32 = LowParse.Bytes32\nmodule U32 = FStar.UInt32\n\nPartial implementation:\nlet parse32_fldata\n (#k: parser_kind)\n (#t: Type)\n (#p: parser k t)\n (p32: parser32 p)\n (sz: nat)\n (sz32: U32.t{U32.v sz32 == sz})\n : Tot (parser32 (parse_fldata p sz)) =\n \n"
},
{
"role": "assistant",
"content": "(fun (input: bytes32) ->\n ((if U32.lt (B32.len input) sz32\n then None\n else\n match p32 (B32.b32slice input 0ul sz32) with\n | Some (v, consumed) -> if consumed = sz32 then Some (v, consumed) else None\n | None -> None)\n <:\n (res: option (t * U32.t) {parser32_correct (parse_fldata p sz) input res})))"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: f: AlgHeap.tree a labs0 -> v: (_: a -> AlgHeap.tree b labs1) -> h: AlgHeap.handler_ty labs0 b labs1\n -> AlgHeap.tree b labs1\nEffect: Prims.Tot\n\nContext:\nmodule AlgHeap\n\n(* Essentially a copy of AlgForAll but using a heap for the state *)\n\nopen FStar.Tactics.V2\nopen FStar.List.Tot\nopen FStar.Universe\nmodule L = Lattice\nmodule Ghost = FStar.Ghost\nmodule Map = FStar.Map\nmodule T = FStar.Tactics.V2\n\ntype loc = int\ntype state = Map.t loc int\n\ntype empty : Type u#aa =\n\ntype op =\n | Read\n | Write\n | Raise\n | Other of int\n\nassume val other_inp : int -> Type u#0\nlet op_inp : op -> Type u#0 =\n function\n | Read -> unit\n | Write -> state\n | Raise -> exn\n | Other i -> other_inp i\n\nassume val other_out : int -> Type u#0\nlet op_out : op -> Type =\n function\n | Read -> state\n | Write -> unit\n | Raise -> empty\n | Other i -> other_inp i\n\nnoeq\ntype tree0 (a:Type) : Type =\n | Return : a -> tree0 a\n | Op : op:op -> i:(op_inp op) -> k:(op_out op -> tree0 a) -> tree0 a\n\ntype ops = list op\n\nlet rec abides #a (labs:ops) (f : tree0 a) : prop =\n begin match f with\n | Op a i k ->\n mem a labs /\\ (forall o. abides labs (k o))\n | Return _ -> True\n end\n\ntype tree (a:Type u#aa)\n (labs : list u#0 op)\n : Type u#aa\n =\n r:(tree0 a){abides labs r}\n\nlet rec interp_at (l1 l2 : ops) (l : op)\n : Lemma (mem l (l1@l2) == (mem l l1 || mem l l2))\n [SMTPat (mem l (l1@l2))]\n = match l1 with\n | [] -> ()\n | _::l1 -> interp_at l1 l2 l\n\nlet sublist (l1 l2 : ops) = forall x. mem x l1 ==> mem x l2\n\nlet rec sublist_at\n (l1 l2 : ops)\n : Lemma (sublist l1 (l1@l2) /\\ sublist l2 (l1@l2))\n [SMTPatOr [[SMTPat (sublist l1 (l1@l2))];\n [SMTPat (sublist l2 (l1@l2))]]]\n = match l1 with\n | [] -> ()\n | _::l1 -> sublist_at l1 l2\n\nlet sublist_at_self\n (l : ops)\n : Lemma (sublist l (l@l))\n [SMTPat (sublist l (l@l))]\n = ()\n \nlet rec abides_sublist_nopat #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c) /\\ sublist l1 l2)\n (ensures (abides l2) c)\n = match c with\n | Return _ -> ()\n | Op a i k ->\n let sub o : Lemma (abides l2 (k o)) =\n abides_sublist_nopat l1 l2 (k o)\n in\n Classical.forall_intro sub\n\nlet abides_sublist #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c) /\\ sublist l1 l2)\n (ensures (abides l2 c))\n [SMTPat (abides l2 c); SMTPat (sublist l1 l2)]\n = abides_sublist_nopat l1 l2 c\n \nlet abides_at_self #a\n (l : ops)\n (c : tree0 a)\n : Lemma (abides (l@l) c <==> abides l c)\n [SMTPat (abides (l@l) c)]\n = (* Trigger some patterns *)\n assert (sublist l (l@l));\n assert (sublist (l@l) l)\n\nlet abides_app #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c \\/ abides l2 c))\n (ensures (abides (l1@l2) c))\n [SMTPat (abides (l1@l2) c)]\n = sublist_at l1 l2\n\n(* Folding a computation tree *)\nval fold_with (#a #b:_) (#labs : ops)\n (f:tree a labs)\n (v : a -> b)\n (h: (o:op{mem o labs} -> op_inp o -> (op_out o -> b) -> b))\n : b\nlet rec fold_with #a #b #labs f v h =\n match f with\n | Return x -> v x\n | Op act i k ->\n let k' (o : op_out act) : b =\n fold_with #_ #_ #labs (k o) v h\n in\n h act i k'\n\nlet handler_ty_l (o:op) (b:Type) (labs:ops) =\n op_inp o -> (op_out o -> tree b labs) -> tree b labs\n\nlet handler_ty (labs0 : ops) (b:Type) (labs1 : ops) : Type =\n o:op{mem o labs0} -> handler_ty_l o b labs1\n\n(* The most generic handling construct, we use it to implement bind.\nIt is actually just a special case of folding. *)\nval handle_with (#a #b:_) (#labs0 #labs1 : ops)\n (f:tree a labs0)\n (v : a -> tree b labs1)\n (h: handler_ty labs0 b labs1)\n\nPartial implementation:\nlet handle_with f v h =\n \n"
},
{
"role": "assistant",
"content": "fold_with f v h"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: o: AlgHeap.op -> b: Type -> labs: AlgHeap.ops -> Type\nEffect: Prims.Tot\n\nContext:\nmodule AlgHeap\n\n(* Essentially a copy of AlgForAll but using a heap for the state *)\n\nopen FStar.Tactics.V2\nopen FStar.List.Tot\nopen FStar.Universe\nmodule L = Lattice\nmodule Ghost = FStar.Ghost\nmodule Map = FStar.Map\nmodule T = FStar.Tactics.V2\n\ntype loc = int\ntype state = Map.t loc int\n\ntype empty : Type u#aa =\n\ntype op =\n | Read\n | Write\n | Raise\n | Other of int\n\nassume val other_inp : int -> Type u#0\nlet op_inp : op -> Type u#0 =\n function\n | Read -> unit\n | Write -> state\n | Raise -> exn\n | Other i -> other_inp i\n\nassume val other_out : int -> Type u#0\nlet op_out : op -> Type =\n function\n | Read -> state\n | Write -> unit\n | Raise -> empty\n | Other i -> other_inp i\n\nnoeq\ntype tree0 (a:Type) : Type =\n | Return : a -> tree0 a\n | Op : op:op -> i:(op_inp op) -> k:(op_out op -> tree0 a) -> tree0 a\n\ntype ops = list op\n\nlet rec abides #a (labs:ops) (f : tree0 a) : prop =\n begin match f with\n | Op a i k ->\n mem a labs /\\ (forall o. abides labs (k o))\n | Return _ -> True\n end\n\ntype tree (a:Type u#aa)\n (labs : list u#0 op)\n : Type u#aa\n =\n r:(tree0 a){abides labs r}\n\nlet rec interp_at (l1 l2 : ops) (l : op)\n : Lemma (mem l (l1@l2) == (mem l l1 || mem l l2))\n [SMTPat (mem l (l1@l2))]\n = match l1 with\n | [] -> ()\n | _::l1 -> interp_at l1 l2 l\n\nlet sublist (l1 l2 : ops) = forall x. mem x l1 ==> mem x l2\n\nlet rec sublist_at\n (l1 l2 : ops)\n : Lemma (sublist l1 (l1@l2) /\\ sublist l2 (l1@l2))\n [SMTPatOr [[SMTPat (sublist l1 (l1@l2))];\n [SMTPat (sublist l2 (l1@l2))]]]\n = match l1 with\n | [] -> ()\n | _::l1 -> sublist_at l1 l2\n\nlet sublist_at_self\n (l : ops)\n : Lemma (sublist l (l@l))\n [SMTPat (sublist l (l@l))]\n = ()\n \nlet rec abides_sublist_nopat #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c) /\\ sublist l1 l2)\n (ensures (abides l2) c)\n = match c with\n | Return _ -> ()\n | Op a i k ->\n let sub o : Lemma (abides l2 (k o)) =\n abides_sublist_nopat l1 l2 (k o)\n in\n Classical.forall_intro sub\n\nlet abides_sublist #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c) /\\ sublist l1 l2)\n (ensures (abides l2 c))\n [SMTPat (abides l2 c); SMTPat (sublist l1 l2)]\n = abides_sublist_nopat l1 l2 c\n \nlet abides_at_self #a\n (l : ops)\n (c : tree0 a)\n : Lemma (abides (l@l) c <==> abides l c)\n [SMTPat (abides (l@l) c)]\n = (* Trigger some patterns *)\n assert (sublist l (l@l));\n assert (sublist (l@l) l)\n\nlet abides_app #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c \\/ abides l2 c))\n (ensures (abides (l1@l2) c))\n [SMTPat (abides (l1@l2) c)]\n = sublist_at l1 l2\n\n(* Folding a computation tree *)\nval fold_with (#a #b:_) (#labs : ops)\n (f:tree a labs)\n (v : a -> b)\n (h: (o:op{mem o labs} -> op_inp o -> (op_out o -> b) -> b))\n : b\nlet rec fold_with #a #b #labs f v h =\n match f with\n | Return x -> v x\n | Op act i k ->\n let k' (o : op_out act) : b =\n fold_with #_ #_ #labs (k o) v h\n in\n h act i k'\n\nPartial implementation:\nlet handler_ty_l (o: op) (b: Type) (labs: ops) =\n \n"
},
{
"role": "assistant",
"content": "op_inp o -> (op_out o -> tree b labs) -> tree b labs"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: d: Pulse.Typing.st_typing g t c -> post_hint: Pulse.Typing.post_hint_opt g\n -> FStar.Tactics.Effect.Tac\n (FStar.Pervasives.dtuple3 Pulse.Syntax.Base.st_term\n (fun _ -> Pulse.Syntax.Base.comp_st)\n (fun t' c' -> Pulse.Typing.st_typing g t' c'))\nEffect: FStar.Tactics.Effect.Tac\n\nContext:\n(*\n Copyright 2023 Microsoft Research\n\n Licensed under the Apache License, Version 2.0 (the \"License\");\n you may not use this file except in compliance with the License.\n You may obtain a copy of the License at\n\n http://www.apache.org/licenses/LICENSE-2.0\n\n Unless required by applicable law or agreed to in writing, software\n distributed under the License is distributed on an \"AS IS\" BASIS,\n WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n See the License for the specific language governing permissions and\n limitations under the License.\n*)\n\nmodule Pulse.Checker.Base\n\nmodule T = FStar.Tactics.V2\nmodule RT = FStar.Reflection.Typing\nmodule Metatheory = Pulse.Typing.Metatheory\nmodule CP = Pulse.Checker.Pure\nmodule RU = Pulse.RuntimeUtils\nmodule FV = Pulse.Typing.FV\nmodule P = Pulse.Syntax.Printer\n\nopen Pulse.Typing.Combinators\nopen Pulse.Typing.Metatheory\n\nlet debug (g:env) (f: unit -> T.Tac string) : T.Tac unit =\n if RU.debug_at_level (fstar_env g) \"pulse.checker\" then\n T.print (f())\n\nlet format_failed_goal (g:env) (ctxt:list term) (goal:list term) =\n let terms_to_strings (ts:list term)= T.map Pulse.Syntax.Printer.term_to_string ts in\n let numbered_list ss = \n let _, s = T.fold_left (fun (i, acc) s -> (i+1, Printf.sprintf \"%d. %s\" i s :: acc)) (1, []) ss in\n String.concat \"\\n \" (List.rev s)\n in\n let format_terms (ts:list term) = numbered_list (terms_to_strings ts) in\n Printf.sprintf \n \"Failed to prove the following goals:\\n \\\n %s\\n\\\n The remaining conjuncts in the separation logic context available for use are:\\n \\\n %s\\n\\\n The typing context is:\\n \\\n %s\\n\"\n (format_terms goal)\n (format_terms ctxt)\n (env_to_string g)\n\n\nlet mk_arrow ty t = RT.mk_arrow (elab_term ty) T.Q_Explicit (elab_term t)\nlet mk_abs ty t = RT.(mk_abs (elab_term ty) T.Q_Explicit (elab_term t))\n\nlet intro_comp_typing (g:env) \n (c:comp_st)\n (pre_typing:tot_typing g (comp_pre c) tm_vprop)\n (i_typing:effect_annot_typing g (effect_annot_of_comp c))\n (res_typing:universe_of g (comp_res c) (comp_u c))\n (x:var { fresh_wrt x g (freevars (comp_post c)) })\n (post_typing:tot_typing (push_binding g x ppname_default (comp_res c)) (open_term (comp_post c) x) tm_vprop)\n : T.Tac (comp_typing g c (universe_of_comp c))\n = let intro_st_comp_typing (st:st_comp { comp_u c == st.u /\\\n comp_pre c == st.pre /\\\n comp_res c == st.res /\\\n comp_post c == st.post } )\n : T.Tac (st_comp_typing g st)\n = STC g st x res_typing pre_typing post_typing\n in\n match c with\n | C_ST st -> \n let stc = intro_st_comp_typing st in\n CT_ST _ _ stc\n | C_STAtomic i obs st -> \n let stc = intro_st_comp_typing st in\n CT_STAtomic _ i obs _ i_typing stc\n | C_STGhost st -> \n let stc = intro_st_comp_typing st in\n CT_STGhost _ _ stc\n\nirreducible\nlet post_typing_as_abstraction\n (#g:env) (#x:var) (#ty:term) (#t:term { fresh_wrt x g (freevars t) })\n (_:tot_typing (push_binding g x ppname_default ty) (open_term t x) tm_vprop)\n : FStar.Ghost.erased (RT.tot_typing (elab_env g) (mk_abs ty t) (mk_arrow ty tm_vprop)) \n = admit()\n\nlet check_effect_annot (g:env) (e:effect_annot) : T.Tac (effect_annot_typing g e) =\n match e with\n | EffectAnnotSTT\n | EffectAnnotGhost -> ()\n | EffectAnnotAtomic { opens } ->\n CP.core_check_term g opens T.E_Total tm_inames\n\nlet intro_post_hint g effect_annot ret_ty_opt post =\n let x = fresh g in\n let ret_ty = \n match ret_ty_opt with\n | None -> tm_fstar RT.unit_ty FStar.Range.range_0\n | Some t -> t\n in\n let ret_ty, _ = CP.instantiate_term_implicits g ret_ty in\n let (| u, ty_typing |) = CP.check_universe g ret_ty in\n let (| post, post_typing |) = CP.check_vprop (push_binding g x ppname_default ret_ty) (open_term_nv post (v_as_nv x)) in \n let post' = close_term post x in\n Pulse.Typing.FV.freevars_close_term post x 0;\n let effect_annot_typing = check_effect_annot g effect_annot in\n assume (open_term post' x == post);\n { g;\n effect_annot;\n effect_annot_typing;\n ret_ty; u; ty_typing;\n post=post';\n x; post_typing_src=post_typing;\n post_typing=post_typing_as_abstraction #_ #_ #_ #post' post_typing }\n\n\nlet comp_typing_as_effect_annot_typing (#g:env) (#c:comp_st) (ct:comp_typing_u g c)\n: effect_annot_typing g (effect_annot_of_comp c)\n= let _, iname_typing = Metatheory.comp_typing_inversion ct in\n match c with\n | C_ST _ -> ()\n | C_STGhost _ -> ()\n | C_STAtomic opens obs _ -> iname_typing\n \n\nlet post_hint_from_comp_typing #g #c ct = \n let st_comp_typing, _ = Metatheory.comp_typing_inversion ct in\n let (| ty_typing, pre_typing, x, post_typing |) = Metatheory.st_comp_typing_inversion st_comp_typing in\n let effect_annot_typing = comp_typing_as_effect_annot_typing ct in\n let p : post_hint_t = \n { g;\n effect_annot=_;\n effect_annot_typing;\n ret_ty = comp_res c; u=comp_u c; \n ty_typing=ty_typing;\n post=comp_post c;\n x;\n post_typing_src=post_typing;\n post_typing=post_typing_as_abstraction post_typing }\n in\n p\n\nlet comp_typing_from_post_hint\n (#g: env)\n (c: comp_st)\n (pre_typing: tot_typing g (comp_pre c) tm_vprop)\n (p:post_hint_for_env g { comp_post_matches_hint c (Some p) })\n: T.Tac (comp_typing_u g c)\n= let x = fresh g in\n if x `Set.mem` freevars p.post //exclude this\n then fail g None \"Impossible: unexpected freevar in post, please file a bug-report\"\n else let post_typing = post_hint_typing g p x in\n intro_comp_typing g c pre_typing\n post_typing.effect_annot_typing\n post_typing.ty_typing \n x post_typing.post_typing\n\n\nlet extend_post_hint g p x tx conjunct conjunct_typing =\n let g' = push_binding g x ppname_default tx in\n let y = fresh g' in\n let g'' = push_binding g' y ppname_default p.ret_ty in\n let p_post_typing_src\n : tot_typing (push_binding p.g p.x ppname_default p.ret_ty)\n (open_term p.post p.x) tm_vprop\n = p.post_typing_src\n in\n let p_post_typing_src''\n : tot_typing g'' (open_term p.post y) tm_vprop\n = RU.magic () //weaken, rename\n in\n let conjunct_typing'\n : tot_typing g' conjunct tm_vprop\n = conjunct_typing\n in\n let conjunct_typing''\n : tot_typing g'' (open_term conjunct y) tm_vprop\n = RU.magic () //weaken\n in\n let new_post = tm_star p.post conjunct in\n let new_post_typing\n : tot_typing g'' (open_term new_post y) tm_vprop\n = Pulse.Typing.star_typing p_post_typing_src'' conjunct_typing''\n in\n assume (fresh_wrt y g'' (freevars new_post));\n let new_post_abs_typing\n : Ghost.erased (RT.tot_typing (elab_env g'') (mk_abs p.ret_ty new_post) (mk_arrow p.ret_ty tm_vprop))\n = post_typing_as_abstraction new_post_typing\n in\n { p with\n g=g';\n post=new_post;\n x=y;\n post_typing_src=new_post_typing;\n post_typing=new_post_abs_typing }\n\nlet k_elab_unit (g:env) (ctxt:term)\n : continuation_elaborator g ctxt g ctxt\n = fun p r -> r\n\nlet k_elab_trans\n (#g0:env) (#g1:env { g1 `env_extends` g0 }) (#g2:env { g2 `env_extends` g1 }) (#ctxt0 #ctxt1 #ctxt2:term)\n (k0:continuation_elaborator g0 ctxt0 g1 ctxt1)\n (k1:continuation_elaborator g1 ctxt1 g2 ctxt2 { g1 `env_extends` g0})\n : continuation_elaborator g0 ctxt0 g2 ctxt2\n = fun post_hint res -> k0 post_hint (k1 post_hint res)\n\nlet comp_st_with_post (c:comp_st) (post:term)\n : c':comp_st { st_comp_of_comp c' == ({ st_comp_of_comp c with post} <: st_comp) } =\n match c with\n | C_ST st -> C_ST { st with post }\n | C_STGhost st -> C_STGhost { st with post }\n | C_STAtomic i obs st -> C_STAtomic i obs {st with post}\n\nlet ve_unit_r g (p:term) : vprop_equiv g (tm_star p tm_emp) p = \n VE_Trans _ _ _ _ (VE_Comm _ _ _) (VE_Unit _ _)\n\nlet st_equiv_trans (#g:env) (#c0 #c1 #c2:comp) (d01:st_equiv g c0 c1) (d12:st_equiv g c1 c2)\n : option (st_equiv g c0 c2)\n = \n match d01 with\n | ST_VPropEquiv _f _c0 _c1 x c0_pre_typing c0_res_typing c0_post_typing eq_res_01 eq_pre_01 eq_post_01 -> (\n let ST_VPropEquiv _f _c1 _c2 y c1_pre_typing c1_res_typing c1_post_typing eq_res_12 eq_pre_12 eq_post_12 = d12 in\n if x = y && eq_tm (comp_res c0) (comp_res c1)\n then Some (\n ST_VPropEquiv g c0 c2 x c0_pre_typing c0_res_typing c0_post_typing\n (RT.Rel_trans _ _ _ _ _ eq_res_01 eq_res_12)\n (VE_Trans _ _ _ _ eq_pre_01 eq_pre_12)\n (VE_Trans _ _ _ _ eq_post_01 eq_post_12)\n )\n else None\n )\n | ST_TotEquiv g t1 t2 u typing eq ->\n let ST_TotEquiv _g _t1 t3 _ _ eq' = d12 in\n let eq'' = Ghost.hide (RT.Rel_trans _ _ _ _ _ eq eq') in\n Some (ST_TotEquiv g t1 t3 u typing eq'')\n\nlet t_equiv #g #st #c (d:st_typing g st c) (#c':comp) (eq:st_equiv g c c')\n : st_typing g st c'\n = match d with\n | T_Equiv _ _ _ _ d0 eq' -> (\n match st_equiv_trans eq' eq with\n | None -> T_Equiv _ _ _ _ d eq\n | Some eq'' -> T_Equiv _ _ _ _ d0 eq''\n )\n | _ -> T_Equiv _ _ _ _ d eq\n\nlet st_equiv_post (#g:env) (#t:st_term) (#c:comp_st) (d:st_typing g t c)\n (post:term { freevars post `Set.subset` freevars (comp_post c)})\n (veq: (x:var { fresh_wrt x g (freevars (comp_post c)) } ->\n vprop_equiv (push_binding g x ppname_default (comp_res c)) \n (open_term (comp_post c) x)\n (open_term post x)))\n : st_typing g t (comp_st_with_post c post)\n = if eq_tm post (comp_post c) then d\n else\n let c' = comp_st_with_post c post in\n let (| u_of, pre_typing, x, post_typing |) = Metatheory.(st_comp_typing_inversion (fst (comp_typing_inversion (st_typing_correctness d)))) in\n let veq = veq x in\n let st_equiv : st_equiv g c c' =\n ST_VPropEquiv g c c' x pre_typing u_of post_typing (RT.Rel_refl _ _ _) (VE_Refl _ _) veq\n in\n t_equiv d st_equiv\n\nlet simplify_post (#g:env) (#t:st_term) (#c:comp_st) (d:st_typing g t c)\n (post:term { comp_post c == tm_star post tm_emp})\n : st_typing g t (comp_st_with_post c post)\n = st_equiv_post d post (fun x -> ve_unit_r (push_binding g x ppname_default (comp_res c)) (open_term post x))\n\nlet simplify_lemma (c:comp_st) (c':comp_st) (post_hint:option post_hint_t)\n : Lemma\n (requires\n comp_post_matches_hint c post_hint /\\\n effect_annot_of_comp c == effect_annot_of_comp c' /\\\n comp_res c' == comp_res c /\\\n comp_u c' == comp_u c /\\\n comp_post c' == tm_star (comp_post c) tm_emp)\n (ensures comp_post_matches_hint (comp_st_with_post c' (comp_post c)) post_hint /\\\n comp_pre (comp_st_with_post c' (comp_post c)) == comp_pre c')\n = ()\n\nlet vprop_equiv_typing_bk (#g:env) (#ctxt:_) (ctxt_typing:tot_typing g ctxt tm_vprop)\n (#p:_) (d:vprop_equiv g p ctxt)\n : tot_typing g p tm_vprop \n = let _, bk = vprop_equiv_typing d in\n bk ctxt_typing\n\nlet comp_with_pre (c:comp_st) (pre:term) =\n match c with\n | C_ST st -> C_ST { st with pre }\n | C_STGhost st -> C_STGhost { st with pre }\n | C_STAtomic i obs st -> C_STAtomic i obs {st with pre}\n\n\nlet st_equiv_pre (#g:env) (#t:st_term) (#c:comp_st) (d:st_typing g t c)\n (pre:term)\n (veq: vprop_equiv g (comp_pre c) pre)\n : st_typing g t (comp_with_pre c pre)\n = if eq_tm pre (comp_pre c) then d\n else\n let c' = comp_with_pre c pre in\n let (| u_of, pre_typing, x, post_typing |) =\n Metatheory.(st_comp_typing_inversion (fst (comp_typing_inversion (st_typing_correctness d)))) in\n let st_equiv : st_equiv g c c' =\n ST_VPropEquiv g c c' x pre_typing u_of post_typing (RT.Rel_refl _ _ _) veq (VE_Refl _ _)\n in\n t_equiv d st_equiv\n\n\n#push-options \"--z3rlimit_factor 4 --ifuel 2 --fuel 0\"\nlet k_elab_equiv_continuation (#g1:env) (#g2:env { g2 `env_extends` g1 }) (#ctxt #ctxt1 #ctxt2:term)\n (k:continuation_elaborator g1 ctxt g2 ctxt1)\n (d:vprop_equiv g2 ctxt1 ctxt2)\n : continuation_elaborator g1 ctxt g2 ctxt2 =\n fun post_hint res ->\n let (| st, c, st_d |) = res in\n let st_d : st_typing g2 st c = st_d in\n assert (comp_pre c == ctxt2);\n let st_d' : st_typing g2 st (comp_with_pre c ctxt1) = st_equiv_pre st_d _ (VE_Sym _ _ _ d) in\n k post_hint (| st, _, st_d' |)\n#pop-options\n\nlet vprop_equiv_typing_fwd (#g:env) (#ctxt:_) (ctxt_typing:tot_typing g ctxt tm_vprop)\n (#p:_) (d:vprop_equiv g ctxt p)\n : tot_typing g p tm_vprop \n = let fwd, _ = vprop_equiv_typing d in\n fwd ctxt_typing\n\n#push-options \"--z3rlimit_factor 4 --ifuel 1 --fuel 0\"\nlet k_elab_equiv_prefix\n (#g1:env) (#g2:env { g2 `env_extends` g1 }) (#ctxt1 #ctxt2 #ctxt:term)\n (k:continuation_elaborator g1 ctxt1 g2 ctxt)\n (d:vprop_equiv g1 ctxt1 ctxt2)\n : continuation_elaborator g1 ctxt2 g2 ctxt =\n fun post_hint res ->\n let framing_token : frame_for_req_in_ctxt g1 ctxt2 ctxt1 = \n let d = VE_Trans _ _ _ _ (VE_Comm _ _ _) (VE_Trans _ _ _ _ (VE_Unit _ _) d) in\n (| tm_emp, emp_typing, d |)\n in\n let res = k post_hint res in\n let (| st, c, st_d |) = res in\n assert (comp_pre c == ctxt1);\n (| _, _, st_equiv_pre st_d _ d |)\n #pop-options\n\nlet k_elab_equiv\n (#g1:env) (#g2:env { g2 `env_extends` g1 }) (#ctxt1 #ctxt1' #ctxt2 #ctxt2':term) \n (k:continuation_elaborator g1 ctxt1 g2 ctxt2)\n (d1:vprop_equiv g1 ctxt1 ctxt1')\n (d2:vprop_equiv g2 ctxt2 ctxt2')\n : continuation_elaborator g1 ctxt1' g2 ctxt2' =\n \n let k : continuation_elaborator g1 ctxt1 g2 ctxt2' =\n k_elab_equiv_continuation k d2 in\n let k : continuation_elaborator g1 ctxt1' g2 ctxt2' =\n k_elab_equiv_prefix k d1 in\n k\n\n#push-options \"--query_stats --fuel 2 --ifuel 2 --split_queries no --z3rlimit_factor 20\"\nopen Pulse.PP\nlet continuation_elaborator_with_bind (#g:env) (ctxt:term)\n (#c1:comp{stateful_comp c1})\n (#e1:st_term)\n (e1_typing:st_typing g e1 c1)\n (ctxt_pre1_typing:tot_typing g (tm_star ctxt (comp_pre c1)) tm_vprop)\n (x:nvar { None? (lookup g (snd x)) })\n : T.Tac (continuation_elaborator\n g\n (tm_star ctxt (comp_pre c1))\n (push_binding g (snd x) (fst x) (comp_res c1))\n (tm_star (open_term (comp_post c1) (snd x)) ctxt)) =\n\n let pre1 = comp_pre c1 in\n let res1 = comp_res c1 in\n let post1 = comp_post c1 in\n let ctxt_typing = star_typing_inversion_l ctxt_pre1_typing in\n // let p_prop = Metatheory.pure_typing_inversion pure_typing in\n let v_eq = VE_Comm g ctxt pre1 in\n let framing_token : frame_for_req_in_ctxt g (tm_star ctxt pre1) pre1 = \n (| ctxt, ctxt_typing, VE_Comm g pre1 ctxt |)\n in\n let (| c1, e1_typing |) =\n apply_frame ctxt_pre1_typing e1_typing framing_token in\n let (| u_of_1, pre_typing, _, _ |) =\n Metatheory.(st_comp_typing_inversion (fst <| comp_typing_inversion (st_typing_correctness e1_typing))) in\n let b = res1 in\n let ppname, x = x in\n let g' = push_binding g x ppname b in\n \n let post1_opened = open_term_nv post1 (v_as_nv x) in\n let k : continuation_elaborator g (tm_star ctxt pre1) g' (tm_star post1_opened ctxt) =\n fun post_hint res ->\n let (| e2, c2, e2_typing |) = res in\n assert (comp_post_matches_hint c2 post_hint);\n let e2_typing : st_typing g' e2 c2 = e2_typing in\n let e2_closed = close_st_term e2 x in\n assume (open_st_term e2_closed x == e2);\n assert (comp_pre c1 == (tm_star ctxt pre1));\n assert (comp_post c1 == tm_star post1 ctxt);\n assert (comp_pre c2 == tm_star post1_opened ctxt);\n assert (open_term (comp_post c1) x == tm_star post1_opened (open_term ctxt x));\n // ctxt is well-typed, hence ln\n assume (open_term ctxt x == ctxt);\n assert (open_term (comp_post c1) x == comp_pre c2);\n // we closed e2 with x\n assume (~ (x `Set.mem` freevars_st e2_closed));\n if x `Set.mem` freevars (comp_post c2)\n then fail g' None \"Impossible: freevar clash when constructing continuation elaborator for bind, please file a bug-report\"\n else (\n let t_typing, post_typing =\n Pulse.Typing.Combinators.bind_res_and_post_typing g c2 x post_hint in\n let g = push_context g \"mk_bind\" e1.range in\n // info_doc g None\n // [prefix 4 1 (doc_of_string \"mk_bind e1 = \") (doc_of_string (Pulse.Syntax.Printer.st_term_to_string e1));\n // prefix 4 1 (doc_of_string \"mk_bind c1 = \") (pp #comp c1);\n // prefix 4 1 (doc_of_string \"mk_bind e2 = \") (doc_of_string (Pulse.Syntax.Printer.st_term_to_string e2));\n // prefix 4 1 (doc_of_string \"mk_bind c2 = \") (pp #comp c2)]\n // ;\n let (| e, c, e_typing |) =\n Pulse.Typing.Combinators.mk_bind\n g (tm_star ctxt pre1) \n e1 e2_closed c1 c2 (ppname, x) e1_typing\n u_of_1 \n e2_typing\n t_typing\n post_typing\n (Some? post_hint)\n in\n (| e, c, e_typing |)\n )\n in\n k\n#pop-options\n\nmodule LN = Pulse.Typing.LN\n#push-options \"--z3rlimit_factor 4 --fuel 1 --ifuel 1\"\n\n\nlet st_comp_typing_with_post_hint \n (#g:env) (#ctxt:_)\n (ctxt_typing:tot_typing g ctxt tm_vprop)\n (post_hint:post_hint_opt g { Some? post_hint })\n (c:comp_st { comp_pre c == ctxt /\\ comp_post_matches_hint c post_hint })\n: st_comp_typing g (st_comp_of_comp c)\n= let st = st_comp_of_comp c in\n let Some ph = post_hint in\n let post_typing_src\n : tot_typing (push_binding ph.g ph.x ppname_default ph.ret_ty)\n (open_term ph.post ph.x) tm_vprop\n = ph.post_typing_src\n in\n let x = fresh g in\n assume (fresh_wrt x g (freevars ph.post));\n assume (None? (lookup g ph.x));\n let post_typing_src\n : tot_typing (push_binding ph.g x ppname_default ph.ret_ty)\n (open_term ph.post x) tm_vprop\n = if x = ph.x\n then post_typing_src\n else \n let open Pulse.Typing.Metatheory.Base in\n let tt :\n tot_typing\n (push_binding ph.g x ppname_default ph.ret_ty)\n (subst_term (open_term ph.post ph.x) (renaming ph.x x))\n (subst_term tm_vprop (renaming ph.x x)) =\n tot_typing_renaming1 ph.g ph.x ph.ret_ty (open_term ph.post ph.x) tm_vprop post_typing_src x\n in\n assert (subst_term tm_vprop (renaming ph.x x) == tm_vprop);\n assume (subst_term (open_term ph.post ph.x) (renaming ph.x x) ==\n open_term ph.post x);\n tt\n in\n let post_typing_src\n : tot_typing (push_binding g x ppname_default ph.ret_ty)\n (open_term ph.post x) tm_vprop\n = //weakening: TODO\n RU.magic ()\n in\n let ty_typing : universe_of ph.g st.res st.u = ph.ty_typing in\n let ty_typing : universe_of g st.res st.u =\n Pulse.Typing.Metatheory.tot_typing_weakening_standard ph.g ty_typing g\n in\n assert (st.res == ph.ret_ty);\n assert (st.post == ph.post);\n STC g st x ty_typing ctxt_typing post_typing_src\n\nlet continuation_elaborator_with_bind_fn (#g:env) (#ctxt:term)\n (ctxt_typing:tot_typing g ctxt tm_vprop)\n (#e1:st_term)\n (#c1:comp { C_Tot? c1 })\n (b:binder{b.binder_ty == comp_res c1})\n (e1_typing:st_typing g e1 c1)\n (x:nvar { None? (lookup g (snd x)) })\n: T.Tac (continuation_elaborator\n g ctxt\n (push_binding g (snd x) ppname_default (comp_res c1)) ctxt)\n= let t1 = comp_res c1 in\n assert ((push_binding g (snd x) (fst x) t1) `env_extends` g);\n fun post_hint (| e2, c2, d2 |) ->\n if None? post_hint then T.fail \"bind_fn: expects the post_hint to be set\";\n let ppname, x = x in\n let e2_closed = close_st_term e2 x in\n assume (open_st_term (close_st_term e2 x) x == e2);\n let e = wr c2 (Tm_Bind {binder=b; head=e1; body=e2_closed}) in\n let (| u, c1_typing |) = Pulse.Typing.Metatheory.Base.st_typing_correctness_ctot e1_typing in\n let c2_typing : comp_typing g c2 (universe_of_comp c2) =\n match c2 with\n | C_ST st -> \n let stc = st_comp_typing_with_post_hint ctxt_typing post_hint c2 in\n CT_ST _ _ stc\n \n | C_STAtomic i obs st -> \n let stc = st_comp_typing_with_post_hint ctxt_typing post_hint c2 in\n let i_typing = CP.core_check_term g i T.E_Total tm_inames in\n CT_STAtomic _ _ obs _ i_typing stc\n\n | C_STGhost st -> \n let stc = st_comp_typing_with_post_hint ctxt_typing post_hint c2 in\n CT_STGhost _ _ stc\n in\n let d : st_typing g e c2 =\n T_BindFn g e1 e2_closed c1 c2 b x e1_typing u c1_typing d2 c2_typing\n in\n (| e, c2, d |)\n\nlet rec check_equiv_emp (g:env) (vp:term)\n : option (vprop_equiv g vp tm_emp)\n = match vp.t with\n | Tm_Emp -> Some (VE_Refl _ _)\n | Tm_Star vp1 vp2 ->\n (match check_equiv_emp g vp1, check_equiv_emp g vp2 with\n | Some d1, Some d2 ->\n let d3 : vprop_equiv g (tm_star vp1 vp2) (tm_star tm_emp tm_emp)\n = VE_Ctxt _ _ _ _ _ d1 d2 in\n let d4 : vprop_equiv g (tm_star tm_emp tm_emp) tm_emp =\n VE_Unit _ _ in\n Some (VE_Trans _ _ _ _ d3 d4)\n | _, _ -> None)\n | _ -> None\n\n\n\nlet emp_inames_included (g:env) (i:term) (_:tot_typing g i tm_inames)\n: prop_validity g (tm_inames_subset tm_emp_inames i)\n= RU.magic()\n\nlet return_in_ctxt (g:env) (y:var) (y_ppname:ppname) (u:universe) (ty:term) (ctxt:vprop)\n (ty_typing:universe_of g ty u)\n (post_hint0:post_hint_opt g { Some? post_hint0 /\\ checker_res_matches_post_hint g post_hint0 y ty ctxt})\n: Pure (st_typing_in_ctxt g ctxt post_hint0)\n (requires lookup g y == Some ty)\n (ensures fun _ -> True)\n= let Some post_hint = post_hint0 in\n let x = fresh g in\n assume (~ (x `Set.mem` freevars post_hint.post));\n let ctag =\n match post_hint.effect_annot with\n | EffectAnnotAtomic _ -> STT_Atomic\n | EffectAnnotGhost -> STT_Ghost\n | _ -> STT\n in\n let y_tm = tm_var {nm_index=y;nm_ppname=y_ppname} in\n let d = T_Return g ctag false u ty y_tm post_hint.post x ty_typing\n (RU.magic ()) // that null_var y is well typed at ty in g, we know since lookup g y == Some ty\n (RU.magic ()) // typing of (open post x) in (g, x) ... post_hint is well-typed, so should get\n in\n let t = wtag (Some ctag) (Tm_Return {expected_type=tm_unknown;insert_eq=false;term=y_tm}) in\n let c = comp_return ctag false u ty y_tm post_hint.post x in\n let d : st_typing g t c = d in\n assume (comp_u c == post_hint.u); // this u should follow from equality of t\n match c, post_hint.effect_annot with\n | C_STAtomic _ obs _, EffectAnnotAtomic { opens } ->\n assert (comp_inames c == tm_emp_inames);\n let pht = post_hint_typing g post_hint x in\n let validity = emp_inames_included g opens pht.effect_annot_typing in\n let d = T_Sub _ _ _ _ d (STS_AtomicInvs _ (st_comp_of_comp c) tm_emp_inames opens obs obs validity) in\n (| _, _, d |)\n | _ -> \n (| _, _, d |)\n\nPartial implementation:\nlet match_comp_res_with_post_hint\n (#g: env)\n (#t: st_term)\n (#c: comp_st)\n (d: st_typing g t c)\n (post_hint: post_hint_opt g)\n : T.Tac (t': st_term & c': comp_st & st_typing g t' c') =\n \n"
},
{
"role": "assistant",
"content": "match post_hint with\n| None -> (| t, c, d |)\n| Some { ret_ty = ret_ty } ->\n let cres = comp_res c in\n if eq_tm cres ret_ty\n then (| t, c, d |)\n else\n match Pulse.Checker.Pure.check_equiv g cres ret_ty with\n | None ->\n fail g\n (Some t.range)\n (Printf.sprintf \"Could not prove equiv for computed type %s and expected type %s\"\n (P.term_to_string cres)\n (P.term_to_string ret_ty))\n | Some tok ->\n let d_equiv:RT.equiv _ (elab_term cres) (elab_term ret_ty) =\n RT.Rel_eq_token _ _ _ (FStar.Squash.return_squash tok)\n in\n let c' = with_st_comp c ({ (st_comp_of_comp c) with res = ret_ty }) in\n let (| cres_typing , cpre_typing , x , cpost_typing |) =\n st_comp_typing_inversion (fst <| comp_typing_inversion (st_typing_correctness d))\n in\n let d_stequiv:st_equiv g c c' =\n ST_VPropEquiv _\n c\n c'\n _\n cpre_typing\n cres_typing\n cpost_typing\n d_equiv\n (VE_Refl _ _)\n (VE_Refl _ _)\n in\n (| t, c', T_Equiv _ _ _ _ d d_stequiv |)"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: l1: AlgHeap.ops -> l2: AlgHeap.ops -> c: AlgHeap.tree0 a\n -> FStar.Pervasives.Lemma (requires AlgHeap.abides l1 c /\\ AlgHeap.sublist l1 l2)\n (ensures AlgHeap.abides l2 c)\n [SMTPat (AlgHeap.abides l2 c); SMTPat (AlgHeap.sublist l1 l2)]\nEffect: FStar.Pervasives.Lemma\n\nContext:\nmodule AlgHeap\n\n(* Essentially a copy of AlgForAll but using a heap for the state *)\n\nopen FStar.Tactics.V2\nopen FStar.List.Tot\nopen FStar.Universe\nmodule L = Lattice\nmodule Ghost = FStar.Ghost\nmodule Map = FStar.Map\nmodule T = FStar.Tactics.V2\n\ntype loc = int\ntype state = Map.t loc int\n\ntype empty : Type u#aa =\n\ntype op =\n | Read\n | Write\n | Raise\n | Other of int\n\nassume val other_inp : int -> Type u#0\nlet op_inp : op -> Type u#0 =\n function\n | Read -> unit\n | Write -> state\n | Raise -> exn\n | Other i -> other_inp i\n\nassume val other_out : int -> Type u#0\nlet op_out : op -> Type =\n function\n | Read -> state\n | Write -> unit\n | Raise -> empty\n | Other i -> other_inp i\n\nnoeq\ntype tree0 (a:Type) : Type =\n | Return : a -> tree0 a\n | Op : op:op -> i:(op_inp op) -> k:(op_out op -> tree0 a) -> tree0 a\n\ntype ops = list op\n\nlet rec abides #a (labs:ops) (f : tree0 a) : prop =\n begin match f with\n | Op a i k ->\n mem a labs /\\ (forall o. abides labs (k o))\n | Return _ -> True\n end\n\ntype tree (a:Type u#aa)\n (labs : list u#0 op)\n : Type u#aa\n =\n r:(tree0 a){abides labs r}\n\nlet rec interp_at (l1 l2 : ops) (l : op)\n : Lemma (mem l (l1@l2) == (mem l l1 || mem l l2))\n [SMTPat (mem l (l1@l2))]\n = match l1 with\n | [] -> ()\n | _::l1 -> interp_at l1 l2 l\n\nlet sublist (l1 l2 : ops) = forall x. mem x l1 ==> mem x l2\n\nlet rec sublist_at\n (l1 l2 : ops)\n : Lemma (sublist l1 (l1@l2) /\\ sublist l2 (l1@l2))\n [SMTPatOr [[SMTPat (sublist l1 (l1@l2))];\n [SMTPat (sublist l2 (l1@l2))]]]\n = match l1 with\n | [] -> ()\n | _::l1 -> sublist_at l1 l2\n\nlet sublist_at_self\n (l : ops)\n : Lemma (sublist l (l@l))\n [SMTPat (sublist l (l@l))]\n = ()\n \nlet rec abides_sublist_nopat #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c) /\\ sublist l1 l2)\n (ensures (abides l2) c)\n = match c with\n | Return _ -> ()\n | Op a i k ->\n let sub o : Lemma (abides l2 (k o)) =\n abides_sublist_nopat l1 l2 (k o)\n in\n Classical.forall_intro sub\n\nPartial implementation:\nlet abides_sublist #a (l1: ops) (l2: ops) (c: tree0 a)\n : Lemma (requires (abides l1 c) /\\ sublist l1 l2)\n (ensures (abides l2 c))\n [SMTPat (abides l2 c); SMTPat (sublist l1 l2)] =\n \n"
},
{
"role": "assistant",
"content": "abides_sublist_nopat l1 l2 c"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n va_mods: Vale.X64.QuickCode.va_mods_t ->\n alg: Vale.AES.AES_common_s.algorithm ->\n offset: Prims.int ->\n auth_b: Vale.X64.Memory.buffer128 ->\n abytes_b: Vale.X64.Memory.buffer128 ->\n in128x6_b: Vale.X64.Memory.buffer128 ->\n out128x6_b: Vale.X64.Memory.buffer128 ->\n in128_b: Vale.X64.Memory.buffer128 ->\n out128_b: Vale.X64.Memory.buffer128 ->\n inout_b: Vale.X64.Memory.buffer128 ->\n iv_b: Vale.X64.Memory.buffer128 ->\n iv: Vale.AES.GCM_s.supported_iv_LE ->\n scratch_b: Vale.X64.Memory.buffer128 ->\n key: FStar.Seq.Base.seq Vale.X64.Memory.nat32 ->\n round_keys: FStar.Seq.Base.seq Vale.X64.Decls.quad32 ->\n keys_b: Vale.X64.Memory.buffer128 ->\n hkeys_b: Vale.X64.Memory.buffer128 ->\n expected_tag: FStar.Seq.Base.seq Vale.X64.Memory.nat8\n -> Vale.X64.QuickCode.va_quickCode Prims.unit\n (Vale.AES.X64.GCMdecryptOpt.va_code_Gcm_blocks_wrapped alg offset)\nEffect: Prims.Tot\n\nContext:\nmodule Vale.AES.X64.GCMdecryptOpt\nopen Vale.Def.Prop_s\nopen Vale.Def.Opaque_s\nopen FStar.Seq\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Seq_s\nopen Vale.Def.Types_s\nopen Vale.Arch.Types\nopen Vale.Arch.HeapImpl\nopen Vale.AES.AES_s\nopen Vale.AES.GCTR_s\nopen Vale.AES.GCTR\nopen Vale.AES.GCM\nopen Vale.AES.GHash_s\nopen Vale.AES.GHash\nopen Vale.AES.GCM_s\nopen Vale.AES.X64.AES\nopen Vale.AES.GF128_s\nopen Vale.AES.GF128\nopen Vale.Poly1305.Math\nopen Vale.AES.GCM_helpers\nopen Vale.AES.X64.GHash\nopen Vale.AES.X64.GCTR\nopen Vale.X64.Machine_s\nopen Vale.X64.Memory\nopen Vale.X64.Stack_i\nopen Vale.X64.State\nopen Vale.X64.Decls\nopen Vale.X64.InsBasic\nopen Vale.X64.InsMem\nopen Vale.X64.InsVector\nopen Vale.X64.InsStack\nopen Vale.X64.InsAes\nopen Vale.X64.QuickCode\nopen Vale.X64.QuickCodes\nopen Vale.AES.X64.GF128_Mul\nopen Vale.X64.Stack\nopen Vale.X64.CPU_Features_s\nopen Vale.Math.Poly2.Bits_s\nopen Vale.AES.X64.AESopt\nopen Vale.AES.X64.AESGCM\nopen Vale.AES.X64.AESopt2\nopen Vale.Lib.Meta\nopen Vale.AES.X64.GCMencryptOpt\nopen Vale.AES.OptPublic\nopen Vale.Lib.Basic\n#reset-options \"--z3rlimit 20 --max_ifuel 0\"\n//-- Gcm_extra_bytes\n\nval va_code_Gcm_extra_bytes : alg:algorithm -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_extra_bytes alg =\n (va_Block (va_CCons (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0)\n (va_op_reg_opr64_reg64 rRax) 0 Secret) (va_CCons (va_code_Mov128 (va_op_xmm_xmm 10)\n (va_op_xmm_xmm 0)) (va_CCons (va_code_Ghash_extra_bytes ()) (va_CCons (va_code_Mov128\n (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_CCons (va_code_Pshufb (va_op_xmm_xmm 0)\n (va_op_xmm_xmm 9)) (va_CCons (va_code_AESEncryptBlock alg) (va_CCons (va_code_Pxor\n (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (va_CCons (va_code_Store128_buffer\n (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax) (va_op_xmm_xmm 10) 0 Secret)\n (va_CNil ()))))))))))\n\nval va_codegen_success_Gcm_extra_bytes : alg:algorithm -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_extra_bytes alg =\n (va_pbool_and (va_codegen_success_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0)\n (va_op_reg_opr64_reg64 rRax) 0 Secret) (va_pbool_and (va_codegen_success_Mov128 (va_op_xmm_xmm\n 10) (va_op_xmm_xmm 0)) (va_pbool_and (va_codegen_success_Ghash_extra_bytes ()) (va_pbool_and\n (va_codegen_success_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_pbool_and\n (va_codegen_success_Pshufb (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (va_pbool_and\n (va_codegen_success_AESEncryptBlock alg) (va_pbool_and (va_codegen_success_Pxor (va_op_xmm_xmm\n 10) (va_op_xmm_xmm 0)) (va_pbool_and (va_codegen_success_Store128_buffer\n (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax) (va_op_xmm_xmm 10) 0 Secret)\n (va_ttrue ())))))))))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_extra_bytes (va_mods:va_mods_t) (alg:algorithm) (inout_b:buffer128) (key:(seq\n nat32)) (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (total_bytes:nat)\n (old_hash:quad32) (completed_quads:(seq quad32)) (h_LE:quad32) : (va_quickCode unit\n (va_code_Gcm_extra_bytes alg)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in let (len:(va_int_range\n 1 1)) = 1 in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 188 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64\n rRax) 0 Secret inout_b 0) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 189 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov128 (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (fun (va_s:va_state) _ -> let\n (hash_input:quad32) = va_get_xmm 0 va_s in va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 193 column 22 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Ghash_extra_bytes hkeys_b total_bytes old_hash h_LE completed_quads) (fun\n (va_s:va_state) _ -> va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 194 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (FStar.Seq.Base.equal #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s)\n inout_b) (FStar.Seq.Base.create #quad32 1 hash_input)) (let (snap:(FStar.Seq.Base.seq\n Vale.X64.Decls.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s) inout_b in va_QSeq\n va_range1\n \"***** PRECONDITION NOT MET AT line 198 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 199 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Pshufb (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (fun (va_s:va_state) _ -> va_QBind\n va_range1\n \"***** PRECONDITION NOT MET AT line 200 column 20 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_AESEncryptBlock alg (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 11 va_s)) key\n round_keys keys_b) (fun (va_s:va_state) _ -> va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 201 column 26 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.AES_s.aes_encrypt_LE_reveal ()) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 204 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Pxor (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 205 column 20 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Store128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax)\n (va_op_xmm_xmm 10) 0 Secret inout_b 0) (fun (va_s:va_state) _ -> va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 207 column 24 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.gctr_partial_reveal ()) (va_QEmpty (()))))))))))))))\n\n\nval va_lemma_Gcm_extra_bytes : va_b0:va_code -> va_s0:va_state -> alg:algorithm ->\n inout_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128 -> total_bytes:nat -> old_hash:quad32 -> completed_quads:(seq quad32) ->\n h_LE:quad32\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_extra_bytes alg) va_s0 /\\ va_get_ok va_s0 /\\ (let\n (len:(va_int_range 1 1)) = 1 in sse_enabled /\\ Vale.X64.Decls.buffers_disjoint128 keys_b\n inout_b /\\ Vale.X64.Decls.buffers_disjoint128 hkeys_b inout_b /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) (va_get_reg64 rRax va_s0) inout_b\n len (va_get_mem_layout va_s0) Secret /\\ len == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 inout_b /\\ va_get_xmm 9 va_s0 == Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\ aes_reqs alg key round_keys\n keys_b (va_get_reg64 rR8 va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\\n pclmulqdq_enabled /\\ Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0\n va_s0) hkeys_b) (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret /\\ va_get_xmm 8 va_s0 == Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.AES.GHash.ghash_incremental0 h_LE old_hash completed_quads) /\\ FStar.Seq.Base.length\n #quad32 completed_quads == total_bytes `op_Division` 16 /\\ total_bytes < 16 `op_Multiply`\n FStar.Seq.Base.length #quad32 completed_quads + 16 /\\ va_get_reg64 rR10 va_s0 == total_bytes\n `op_Modulus` 16 /\\ total_bytes `op_Modulus` 16 =!= 0 /\\ (0 < total_bytes /\\ total_bytes < 16\n `op_Multiply` Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes) /\\ 16 `op_Multiply`\n (Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes - 1) < total_bytes)))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (let (len:(va_int_range 1 1)) = 1 in Vale.X64.Decls.modifies_buffer128 inout_b\n (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5 va_sM) /\\ Vale.AES.GCTR.gctr_partial alg len\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b) key (va_get_xmm 11 va_s0) /\\ (let raw_quads =\n FStar.Seq.Base.append #quad32 completed_quads (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0)\n inout_b) in let input_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads) 0 total_bytes in let padded_bytes =\n Vale.AES.GCTR_s.pad_to_128_bits input_bytes in let input_quads =\n Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in l_and (FStar.Seq.Base.length\n #Vale.Def.Types_s.quad32 input_quads > 0) (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8\n va_sM) == Vale.AES.GHash.ghash_incremental h_LE old_hash input_quads))) /\\ va_state_eq va_sM\n (va_update_flags va_sM (va_update_mem_heaplet 5 va_sM (va_update_xmm 10 va_sM (va_update_xmm 8\n va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5 va_sM (va_update_xmm 4\n va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1 va_sM (va_update_xmm 0\n va_sM (va_update_reg64 rR11 va_sM (va_update_reg64 rRcx va_sM (va_update_ok va_sM\n (va_update_mem va_sM va_s0))))))))))))))))))\n[@\"opaque_to_smt\"]\nlet va_lemma_Gcm_extra_bytes va_b0 va_s0 alg inout_b key round_keys keys_b hkeys_b total_bytes\n old_hash completed_quads h_LE =\n let (va_mods:va_mods_t) = [va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8;\n va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm\n 1; va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_ok; va_Mod_mem] in\n let va_qc = va_qcode_Gcm_extra_bytes va_mods alg inout_b key round_keys keys_b hkeys_b\n total_bytes old_hash completed_quads h_LE in\n let (va_sM, va_fM, va_g) = va_wp_sound_code_norm (va_code_Gcm_extra_bytes alg) va_qc va_s0 (fun\n va_s0 va_sM va_g -> let () = va_g in label va_range1\n \"***** POSTCONDITION NOT MET AT line 121 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\ (let (len:(va_int_range 1 1)) = 1 in label va_range1\n \"***** POSTCONDITION NOT MET AT line 174 column 55 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 177 column 95 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.AES.GCTR.gctr_partial alg len (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM) inout_b) key (va_get_xmm 11 va_s0)) /\\ label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 180 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let raw_quads = FStar.Seq.Base.append #quad32 completed_quads (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 181 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let input_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads) 0 total_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 182 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let padded_bytes = Vale.AES.GCTR_s.pad_to_128_bits input_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 183 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let input_quads = Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 186 column 59 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (l_and (FStar.Seq.Base.length #Vale.Def.Types_s.quad32 input_quads > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE old_hash input_quads)))))))) in\n assert_norm (va_qc.mods == va_mods);\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm\n 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1;\n va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_ok; va_Mod_mem]) va_sM va_s0;\n (va_sM, va_fM)\n\n[@ va_qattr]\nlet va_wp_Gcm_extra_bytes (alg:algorithm) (inout_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (total_bytes:nat) (old_hash:quad32)\n (completed_quads:(seq quad32)) (h_LE:quad32) (va_s0:va_state) (va_k:(va_state -> unit -> Type0))\n : Type0 =\n (va_get_ok va_s0 /\\ (let (len:(va_int_range 1 1)) = 1 in sse_enabled /\\\n Vale.X64.Decls.buffers_disjoint128 keys_b inout_b /\\ Vale.X64.Decls.buffers_disjoint128 hkeys_b\n inout_b /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) (va_get_reg64 rRax\n va_s0) inout_b len (va_get_mem_layout va_s0) Secret /\\ len == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 inout_b /\\ va_get_xmm 9 va_s0 == Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\ aes_reqs alg key round_keys\n keys_b (va_get_reg64 rR8 va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\\n pclmulqdq_enabled /\\ Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0\n va_s0) hkeys_b) (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret /\\ va_get_xmm 8 va_s0 == Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.AES.GHash.ghash_incremental0 h_LE old_hash completed_quads) /\\ FStar.Seq.Base.length\n #quad32 completed_quads == total_bytes `op_Division` 16 /\\ total_bytes < 16 `op_Multiply`\n FStar.Seq.Base.length #quad32 completed_quads + 16 /\\ va_get_reg64 rR10 va_s0 == total_bytes\n `op_Modulus` 16 /\\ total_bytes `op_Modulus` 16 =!= 0 /\\ (0 < total_bytes /\\ total_bytes < 16\n `op_Multiply` Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes) /\\ 16 `op_Multiply`\n (Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes - 1) < total_bytes) /\\ (forall\n (va_x_mem:vale_heap) (va_x_rcx:nat64) (va_x_r11:nat64) (va_x_xmm0:quad32) (va_x_xmm1:quad32)\n (va_x_xmm2:quad32) (va_x_xmm3:quad32) (va_x_xmm4:quad32) (va_x_xmm5:quad32) (va_x_xmm6:quad32)\n (va_x_xmm7:quad32) (va_x_xmm8:quad32) (va_x_xmm10:quad32) (va_x_heap5:vale_heap)\n (va_x_efl:Vale.X64.Flags.t) . let va_sM = va_upd_flags va_x_efl (va_upd_mem_heaplet 5\n va_x_heap5 (va_upd_xmm 10 va_x_xmm10 (va_upd_xmm 8 va_x_xmm8 (va_upd_xmm 7 va_x_xmm7\n (va_upd_xmm 6 va_x_xmm6 (va_upd_xmm 5 va_x_xmm5 (va_upd_xmm 4 va_x_xmm4 (va_upd_xmm 3 va_x_xmm3\n (va_upd_xmm 2 va_x_xmm2 (va_upd_xmm 1 va_x_xmm1 (va_upd_xmm 0 va_x_xmm0 (va_upd_reg64 rR11\n va_x_r11 (va_upd_reg64 rRcx va_x_rcx (va_upd_mem va_x_mem va_s0)))))))))))))) in va_get_ok\n va_sM /\\ (let (len:(va_int_range 1 1)) = 1 in Vale.X64.Decls.modifies_buffer128 inout_b\n (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5 va_sM) /\\ Vale.AES.GCTR.gctr_partial alg len\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b) key (va_get_xmm 11 va_s0) /\\ (let raw_quads =\n FStar.Seq.Base.append #quad32 completed_quads (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0)\n inout_b) in let input_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads) 0 total_bytes in let padded_bytes =\n Vale.AES.GCTR_s.pad_to_128_bits input_bytes in let input_quads =\n Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in l_and (FStar.Seq.Base.length\n #Vale.Def.Types_s.quad32 input_quads > 0) (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8\n va_sM) == Vale.AES.GHash.ghash_incremental h_LE old_hash input_quads))) ==> va_k va_sM (())))\n\n\nval va_wpProof_Gcm_extra_bytes : alg:algorithm -> inout_b:buffer128 -> key:(seq nat32) ->\n round_keys:(seq quad32) -> keys_b:buffer128 -> hkeys_b:buffer128 -> total_bytes:nat ->\n old_hash:quad32 -> completed_quads:(seq quad32) -> h_LE:quad32 -> va_s0:va_state ->\n va_k:(va_state -> unit -> Type0)\n -> Ghost (va_state & va_fuel & unit)\n (requires (va_t_require va_s0 /\\ va_wp_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b\n total_bytes old_hash completed_quads h_LE va_s0 va_k))\n (ensures (fun (va_sM, va_f0, va_g) -> va_t_ensure (va_code_Gcm_extra_bytes alg) ([va_Mod_flags;\n va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5;\n va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR11;\n va_Mod_reg64 rRcx; va_Mod_mem]) va_s0 va_k ((va_sM, va_f0, va_g))))\n[@\"opaque_to_smt\"]\nlet va_wpProof_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b total_bytes old_hash\n completed_quads h_LE va_s0 va_k =\n let (va_sM, va_f0) = va_lemma_Gcm_extra_bytes (va_code_Gcm_extra_bytes alg) va_s0 alg inout_b key\n round_keys keys_b hkeys_b total_bytes old_hash completed_quads h_LE in\n va_lemma_upd_update va_sM;\n assert (va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 5 va_sM (va_update_xmm 10\n va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5\n va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1\n va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR11 va_sM (va_update_reg64 rRcx va_sM\n (va_update_ok va_sM (va_update_mem va_sM va_s0)))))))))))))))));\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm\n 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1;\n va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_mem]) va_sM va_s0;\n let va_g = () in\n (va_sM, va_f0, va_g)\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_quick_Gcm_extra_bytes (alg:algorithm) (inout_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (total_bytes:nat) (old_hash:quad32)\n (completed_quads:(seq quad32)) (h_LE:quad32) : (va_quickCode unit (va_code_Gcm_extra_bytes alg)) =\n (va_QProc (va_code_Gcm_extra_bytes alg) ([va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10;\n va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm\n 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_mem])\n (va_wp_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b total_bytes old_hash\n completed_quads h_LE) (va_wpProof_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b\n total_bytes old_hash completed_quads h_LE))\n//--\n//-- Gcm_blocks128\n\nval va_code_Gcm_blocks128 : alg:algorithm -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_blocks128 alg =\n (va_Block (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRbx) (va_op_opr64_reg64 rRdi))\n (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR12) (va_op_opr64_reg64 rRdx)) (va_CCons\n (va_code_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRax)) (va_CCons\n (va_code_Ghash_buffer ()) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRdi)\n (va_op_opr64_reg64 rRbx)) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRdx)\n (va_op_opr64_reg64 rR12)) (va_CCons (va_code_Gctr_blocks128 alg) (va_CNil ())))))))))\n\nval va_codegen_success_Gcm_blocks128 : alg:algorithm -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_blocks128 alg =\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rRbx) (va_op_opr64_reg64 rRdi))\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR12) (va_op_opr64_reg64 rRdx))\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRax))\n (va_pbool_and (va_codegen_success_Ghash_buffer ()) (va_pbool_and (va_codegen_success_Mov64\n (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRbx)) (va_pbool_and (va_codegen_success_Mov64\n (va_op_dst_opr64_reg64 rRdx) (va_op_opr64_reg64 rR12)) (va_pbool_and\n (va_codegen_success_Gctr_blocks128 alg) (va_ttrue ()))))))))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_blocks128 (va_mods:va_mods_t) (alg:algorithm) (in_b:buffer128) (out_b:buffer128)\n (key:(seq nat32)) (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (h_LE:quad32)\n : (va_quickCode unit (va_code_Gcm_blocks128 alg)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 274 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRbx) (va_op_opr64_reg64 rRdi)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 275 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR12) (va_op_opr64_reg64 rRdx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 276 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRax)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 277 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Ghash_buffer hkeys_b in_b h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8\n va_old_s))) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 278 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRbx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 279 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRdx) (va_op_opr64_reg64 rR12)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 280 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gctr_blocks128 alg in_b out_b key round_keys keys_b) (va_QEmpty (()))))))))))\n\n\nval va_lemma_Gcm_blocks128 : va_b0:va_code -> va_s0:va_state -> alg:algorithm -> in_b:buffer128 ->\n out_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128 -> h_LE:quad32\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_blocks128 alg) va_s0 /\\ va_get_ok va_s0 /\\\n (sse_enabled /\\ Vale.X64.Decls.buffers_disjoint128 keys_b out_b /\\\n Vale.X64.Decls.buffers_disjoint128 hkeys_b out_b /\\ (Vale.X64.Decls.buffers_disjoint128 in_b\n out_b \\/ in_b == out_b) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0)\n (va_get_reg64 rRax va_s0) in_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1 va_s0) (va_get_reg64 rRdi va_s0) out_b\n (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\ va_get_reg64 rRax va_s0 + 16\n `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply`\n va_get_reg64 rRdx va_s0 < pow2_64 /\\ l_and (Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 in_b == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out_b)\n (Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b < pow2_32) /\\ va_get_reg64 rRdx\n va_s0 == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b /\\ va_get_xmm 9 va_s0 ==\n Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\\n va_get_reg64 rRdx va_s0 < pow2_32 /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rR8\n va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret)))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (Vale.X64.Decls.modifies_buffer128 out_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM) /\\ Vale.AES.GCTR.gctr_partial alg (va_get_reg64 rRdx va_s0) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b) key\n (va_get_xmm 11 va_s0) /\\ va_get_xmm 11 va_sM == Vale.AES.GCTR.inc32lite (va_get_xmm 11 va_s0)\n (va_get_reg64 rRdx va_s0) /\\ (va_get_reg64 rRdx va_s0 == 0 ==> l_and (va_get_xmm 8 va_sM ==\n va_get_xmm 8 va_s0) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b ==\n Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) out_b)) /\\ (va_get_reg64 rRdx va_s0 > 0 ==>\n l_and (va_get_reg64 rRdx va_s0 <= FStar.Seq.Base.length #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) ==> FStar.Seq.Base.length\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.slice #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) 0 (va_get_reg64 rRdx va_s0)) > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s0)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b)))) /\\ va_state_eq va_sM (va_update_flags va_sM\n (va_update_mem_heaplet 1 va_sM (va_update_xmm 10 va_sM (va_update_xmm 11 va_sM (va_update_xmm 8\n va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5 va_sM (va_update_xmm 4\n va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1 va_sM (va_update_xmm 0\n va_sM (va_update_reg64 rR12 va_sM (va_update_reg64 rRdx va_sM (va_update_reg64 rR10 va_sM\n (va_update_reg64 rR11 va_sM (va_update_reg64 rRdi va_sM (va_update_reg64 rRbx va_sM\n (va_update_ok va_sM (va_update_mem va_sM va_s0)))))))))))))))))))))))\n[@\"opaque_to_smt\"]\nlet va_lemma_Gcm_blocks128 va_b0 va_s0 alg in_b out_b key round_keys keys_b hkeys_b h_LE =\n let (va_mods:va_mods_t) = [va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11;\n va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm\n 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10;\n va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_ok; va_Mod_mem] in\n let va_qc = va_qcode_Gcm_blocks128 va_mods alg in_b out_b key round_keys keys_b hkeys_b h_LE in\n let (va_sM, va_fM, va_g) = va_wp_sound_code_norm (va_code_Gcm_blocks128 alg) va_qc va_s0 (fun\n va_s0 va_sM va_g -> let () = va_g in label va_range1\n \"***** POSTCONDITION NOT MET AT line 210 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\ (label va_range1\n \"***** POSTCONDITION NOT MET AT line 255 column 53 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 261 column 95 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.AES.GCTR.gctr_partial alg (va_get_reg64 rRdx va_s0) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b) key\n (va_get_xmm 11 va_s0)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 262 column 45 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_xmm 11 va_sM == Vale.AES.GCTR.inc32lite (va_get_xmm 11 va_s0) (va_get_reg64 rRdx\n va_s0)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 265 column 93 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRdx va_s0 == 0 ==> l_and (va_get_xmm 8 va_sM == va_get_xmm 8 va_s0)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b == Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) out_b)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 267 column 131 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRdx va_s0 > 0 ==> l_and (va_get_reg64 rRdx va_s0 <= FStar.Seq.Base.length\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) ==>\n FStar.Seq.Base.length #Vale.X64.Decls.quad32 (FStar.Seq.Base.slice #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) 0 (va_get_reg64 rRdx va_s0)) > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s0)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b))))) in\n assert_norm (va_qc.mods == va_mods);\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11; va_Mod_xmm\n 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2;\n va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10;\n va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_ok; va_Mod_mem]) va_sM va_s0;\n (va_sM, va_fM)\n\n[@ va_qattr]\nlet va_wp_Gcm_blocks128 (alg:algorithm) (in_b:buffer128) (out_b:buffer128) (key:(seq nat32))\n (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (h_LE:quad32) (va_s0:va_state)\n (va_k:(va_state -> unit -> Type0)) : Type0 =\n (va_get_ok va_s0 /\\ (sse_enabled /\\ Vale.X64.Decls.buffers_disjoint128 keys_b out_b /\\\n Vale.X64.Decls.buffers_disjoint128 hkeys_b out_b /\\ (Vale.X64.Decls.buffers_disjoint128 in_b\n out_b \\/ in_b == out_b) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0)\n (va_get_reg64 rRax va_s0) in_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1 va_s0) (va_get_reg64 rRdi va_s0) out_b\n (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\ va_get_reg64 rRax va_s0 + 16\n `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply`\n va_get_reg64 rRdx va_s0 < pow2_64 /\\ l_and (Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 in_b == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out_b)\n (Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b < pow2_32) /\\ va_get_reg64 rRdx\n va_s0 == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b /\\ va_get_xmm 9 va_s0 ==\n Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\\n va_get_reg64 rRdx va_s0 < pow2_32 /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rR8\n va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret) /\\ (forall (va_x_mem:vale_heap) (va_x_rbx:nat64) (va_x_rdi:nat64) (va_x_r11:nat64)\n (va_x_r10:nat64) (va_x_rdx:nat64) (va_x_r12:nat64) (va_x_xmm0:quad32) (va_x_xmm1:quad32)\n (va_x_xmm2:quad32) (va_x_xmm3:quad32) (va_x_xmm4:quad32) (va_x_xmm5:quad32) (va_x_xmm6:quad32)\n (va_x_xmm7:quad32) (va_x_xmm8:quad32) (va_x_xmm11:quad32) (va_x_xmm10:quad32)\n (va_x_heap1:vale_heap) (va_x_efl:Vale.X64.Flags.t) . let va_sM = va_upd_flags va_x_efl\n (va_upd_mem_heaplet 1 va_x_heap1 (va_upd_xmm 10 va_x_xmm10 (va_upd_xmm 11 va_x_xmm11\n (va_upd_xmm 8 va_x_xmm8 (va_upd_xmm 7 va_x_xmm7 (va_upd_xmm 6 va_x_xmm6 (va_upd_xmm 5 va_x_xmm5\n (va_upd_xmm 4 va_x_xmm4 (va_upd_xmm 3 va_x_xmm3 (va_upd_xmm 2 va_x_xmm2 (va_upd_xmm 1 va_x_xmm1\n (va_upd_xmm 0 va_x_xmm0 (va_upd_reg64 rR12 va_x_r12 (va_upd_reg64 rRdx va_x_rdx (va_upd_reg64\n rR10 va_x_r10 (va_upd_reg64 rR11 va_x_r11 (va_upd_reg64 rRdi va_x_rdi (va_upd_reg64 rRbx\n va_x_rbx (va_upd_mem va_x_mem va_s0))))))))))))))))))) in va_get_ok va_sM /\\\n (Vale.X64.Decls.modifies_buffer128 out_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM) /\\ Vale.AES.GCTR.gctr_partial alg (va_get_reg64 rRdx va_s0) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b) key\n (va_get_xmm 11 va_s0) /\\ va_get_xmm 11 va_sM == Vale.AES.GCTR.inc32lite (va_get_xmm 11 va_s0)\n (va_get_reg64 rRdx va_s0) /\\ (va_get_reg64 rRdx va_s0 == 0 ==> l_and (va_get_xmm 8 va_sM ==\n va_get_xmm 8 va_s0) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b ==\n Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) out_b)) /\\ (va_get_reg64 rRdx va_s0 > 0 ==>\n l_and (va_get_reg64 rRdx va_s0 <= FStar.Seq.Base.length #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) ==> FStar.Seq.Base.length\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.slice #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) 0 (va_get_reg64 rRdx va_s0)) > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s0)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b)))) ==> va_k va_sM (())))\n\n\nval va_wpProof_Gcm_blocks128 : alg:algorithm -> in_b:buffer128 -> out_b:buffer128 -> key:(seq\n nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 -> hkeys_b:buffer128 -> h_LE:quad32 ->\n va_s0:va_state -> va_k:(va_state -> unit -> Type0)\n -> Ghost (va_state & va_fuel & unit)\n (requires (va_t_require va_s0 /\\ va_wp_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b\n h_LE va_s0 va_k))\n (ensures (fun (va_sM, va_f0, va_g) -> va_t_ensure (va_code_Gcm_blocks128 alg) ([va_Mod_flags;\n va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6;\n va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0;\n va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10; va_Mod_reg64 rR11; va_Mod_reg64 rRdi;\n va_Mod_reg64 rRbx; va_Mod_mem]) va_s0 va_k ((va_sM, va_f0, va_g))))\n[@\"opaque_to_smt\"]\nlet va_wpProof_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b h_LE va_s0 va_k =\n let (va_sM, va_f0) = va_lemma_Gcm_blocks128 (va_code_Gcm_blocks128 alg) va_s0 alg in_b out_b key\n round_keys keys_b hkeys_b h_LE in\n va_lemma_upd_update va_sM;\n assert (va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 1 va_sM (va_update_xmm 10\n va_sM (va_update_xmm 11 va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6\n va_sM (va_update_xmm 5 va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2\n va_sM (va_update_xmm 1 va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR12 va_sM\n (va_update_reg64 rRdx va_sM (va_update_reg64 rR10 va_sM (va_update_reg64 rR11 va_sM\n (va_update_reg64 rRdi va_sM (va_update_reg64 rRbx va_sM (va_update_ok va_sM (va_update_mem\n va_sM va_s0))))))))))))))))))))));\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11; va_Mod_xmm\n 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2;\n va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10;\n va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_mem]) va_sM va_s0;\n let va_g = () in\n (va_sM, va_f0, va_g)\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_quick_Gcm_blocks128 (alg:algorithm) (in_b:buffer128) (out_b:buffer128) (key:(seq nat32))\n (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (h_LE:quad32) : (va_quickCode\n unit (va_code_Gcm_blocks128 alg)) =\n (va_QProc (va_code_Gcm_blocks128 alg) ([va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10;\n va_Mod_xmm 11; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm\n 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64\n rR10; va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_mem])\n (va_wp_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b h_LE)\n (va_wpProof_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b h_LE))\n//--\n//-- Gcm_blocks\n#push-options \"--z3rlimit 1000\"\n\nval va_code_Gcm_blocks : alg:algorithm -> offset:int -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_blocks alg offset =\n (va_Block (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR13) (va_op_opr64_reg64 rRcx))\n (va_CCons (va_code_AddLea64 (va_op_dst_opr64_reg64 rR9) (va_op_opr64_reg64 rR9) (va_const_opr64\n 32)) (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rRbx) (va_op_reg_opr64_reg64 rRsp)\n (offset + 0)) (va_CCons (va_code_Gcm_blocks_auth ()) (va_CCons (va_code_Load64_stack\n (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 8)) (va_CCons\n (va_code_Load64_stack (va_op_dst_opr64_reg64 rRsi) (va_op_reg_opr64_reg64 rRsp) (offset + 16))\n (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp)\n (offset + 24)) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRcx) (va_op_opr64_reg64 rR13))\n (va_CCons (va_code_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (va_CCons\n (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 2) (va_op_xmm_xmm 1) (va_op_reg_opr64_reg64\n rR8) 0 Public) (va_CCons (va_code_Store128_buffer (va_op_heaplet_mem_heaplet 3)\n (va_op_reg_opr64_reg64 rRbp) (va_op_xmm_xmm 1) 0 Secret) (va_CCons (va_code_Load_one_lsb\n (va_op_xmm_xmm 10)) (va_CCons (va_code_VPaddd (va_op_xmm_xmm 1) (va_op_xmm_xmm 1)\n (va_op_xmm_xmm 10)) (va_CCons (va_code_AES_GCM_decrypt_6mult alg) (va_CCons\n (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 11) (va_op_reg_opr64_reg64\n rRbp) 32 Secret) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR8) (va_op_opr64_reg64 rRcx))\n (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp)\n (offset + 32)) (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rRdi)\n (va_op_reg_opr64_reg64 rRsp) (offset + 40)) (va_CCons (va_code_Load64_stack\n (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp) (offset + 48)) (va_CCons\n (va_code_Mov64 (va_op_dst_opr64_reg64 rR14) (va_op_opr64_reg64 rRdx)) (va_CCons\n (va_code_InitPshufbMask (va_op_xmm_xmm 9) (va_op_reg_opr64_reg64 rR12)) (va_CCons\n (va_code_Pshufb (va_op_xmm_xmm 11) (va_op_xmm_xmm 9)) (va_CCons (va_code_Gcm_blocks128 alg)\n (va_CCons (va_code_Stack_lemma ()) (va_CCons (va_code_Add64 (va_op_dst_opr64_reg64 rR14)\n (va_opr_code_Stack (va_op_reg64_reg64 rRsp) (offset + 24) Public)) (va_CCons (va_code_IMul64\n (va_op_dst_opr64_reg64 rR14) (va_const_opr64 16)) (va_CCons (va_code_Load64_stack\n (va_op_dst_opr64_reg64 rR13) (va_op_reg_opr64_reg64 rRsp) (offset + 64)) (va_CCons (va_IfElse\n (va_cmp_gt (va_op_cmp_reg64 rR13) (va_op_cmp_reg64 rR14)) (va_Block (va_CCons\n (va_code_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp) (offset + 56))\n (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR10) (va_op_opr64_reg64 rR13)) (va_CCons\n (va_code_And64 (va_op_dst_opr64_reg64 rR10) (va_const_opr64 15)) (va_CCons\n (va_code_Gcm_extra_bytes alg) (va_CCons (va_Block (va_CNil ())) (va_CNil ()))))))) (va_Block\n (va_CNil ()))) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR11) (va_op_opr64_reg64 rR15))\n (va_CCons (va_code_Gcm_make_length_quad ()) (va_CCons (va_code_Ghash_register ()) (va_CCons\n (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64\n rRbp) 0 Secret) (va_CCons (va_code_Gctr_register alg) (va_CCons (va_Block (va_CNil ()))\n (va_CNil ()))))))))))))))))))))))))))))))))))))\n\nval va_codegen_success_Gcm_blocks : alg:algorithm -> offset:int -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_blocks alg offset =\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR13) (va_op_opr64_reg64 rRcx))\n (va_pbool_and (va_codegen_success_AddLea64 (va_op_dst_opr64_reg64 rR9) (va_op_opr64_reg64 rR9)\n (va_const_opr64 32)) (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64\n rRbx) (va_op_reg_opr64_reg64 rRsp) (offset + 0)) (va_pbool_and\n (va_codegen_success_Gcm_blocks_auth ()) (va_pbool_and (va_codegen_success_Load64_stack\n (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 8)) (va_pbool_and\n (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRsi) (va_op_reg_opr64_reg64 rRsp)\n (offset + 16)) (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRdx)\n (va_op_reg_opr64_reg64 rRsp) (offset + 24)) (va_pbool_and (va_codegen_success_Mov64\n (va_op_dst_opr64_reg64 rRcx) (va_op_opr64_reg64 rR13)) (va_pbool_and (va_codegen_success_Mov128\n (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (va_pbool_and (va_codegen_success_Load128_buffer\n (va_op_heaplet_mem_heaplet 2) (va_op_xmm_xmm 1) (va_op_reg_opr64_reg64 rR8) 0 Public)\n (va_pbool_and (va_codegen_success_Store128_buffer (va_op_heaplet_mem_heaplet 3)\n (va_op_reg_opr64_reg64 rRbp) (va_op_xmm_xmm 1) 0 Secret) (va_pbool_and\n (va_codegen_success_Load_one_lsb (va_op_xmm_xmm 10)) (va_pbool_and (va_codegen_success_VPaddd\n (va_op_xmm_xmm 1) (va_op_xmm_xmm 1) (va_op_xmm_xmm 10)) (va_pbool_and\n (va_codegen_success_AES_GCM_decrypt_6mult alg) (va_pbool_and (va_codegen_success_Load128_buffer\n (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 11) (va_op_reg_opr64_reg64 rRbp) 32 Secret)\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR8) (va_op_opr64_reg64 rRcx))\n (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRax)\n (va_op_reg_opr64_reg64 rRsp) (offset + 32)) (va_pbool_and (va_codegen_success_Load64_stack\n (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 40)) (va_pbool_and\n (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp)\n (offset + 48)) (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR14)\n (va_op_opr64_reg64 rRdx)) (va_pbool_and (va_codegen_success_InitPshufbMask (va_op_xmm_xmm 9)\n (va_op_reg_opr64_reg64 rR12)) (va_pbool_and (va_codegen_success_Pshufb (va_op_xmm_xmm 11)\n (va_op_xmm_xmm 9)) (va_pbool_and (va_codegen_success_Gcm_blocks128 alg) (va_pbool_and\n (va_codegen_success_Stack_lemma ()) (va_pbool_and (va_codegen_success_Add64\n (va_op_dst_opr64_reg64 rR14) (va_opr_code_Stack (va_op_reg64_reg64 rRsp) (offset + 24) Public))\n (va_pbool_and (va_codegen_success_IMul64 (va_op_dst_opr64_reg64 rR14) (va_const_opr64 16))\n (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rR13)\n (va_op_reg_opr64_reg64 rRsp) (offset + 64)) (va_pbool_and (va_pbool_and\n (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp)\n (offset + 56)) (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR10)\n (va_op_opr64_reg64 rR13)) (va_pbool_and (va_codegen_success_And64 (va_op_dst_opr64_reg64 rR10)\n (va_const_opr64 15)) (va_codegen_success_Gcm_extra_bytes alg)))) (va_pbool_and\n (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR11) (va_op_opr64_reg64 rR15)) (va_pbool_and\n (va_codegen_success_Gcm_make_length_quad ()) (va_pbool_and (va_codegen_success_Ghash_register\n ()) (va_pbool_and (va_codegen_success_Load128_buffer (va_op_heaplet_mem_heaplet 3)\n (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64 rRbp) 0 Secret) (va_pbool_and\n (va_codegen_success_Gctr_register alg) (va_ttrue ()))))))))))))))))))))))))))))))))))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_blocks (va_mods:va_mods_t) (alg:algorithm) (offset:int) (auth_b:buffer128)\n (abytes_b:buffer128) (in128x6_b:buffer128) (out128x6_b:buffer128) (in128_b:buffer128)\n (out128_b:buffer128) (inout_b:buffer128) (iv_b:buffer128) (scratch_b:buffer128) (key:(seq nat32))\n (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) : (va_quickCode unit\n (va_code_Gcm_blocks alg offset)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in let\n (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 0) (va_get_stack va_s) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 8) (va_get_stack va_s) in let\n (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 16) (va_get_stack va_s) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 24) (va_get_stack va_s) in let\n (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 32) (va_get_stack va_s) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 40) (va_get_stack va_s) in let\n (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset\n + 48) (va_get_stack va_s) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 56) (va_get_stack va_s) in let\n (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s\n + offset + 64) (va_get_stack va_s) in let (h_LE:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.X64.Decls.buffer128_read hkeys_b 2\n (va_get_mem_heaplet 0 va_old_s)) in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 463 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR13) (va_op_opr64_reg64 rRcx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 464 column 13 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_AddLea64 (va_op_dst_opr64_reg64 rR9) (va_op_opr64_reg64 rR9) (va_const_opr64 32))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 465 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRbx) (va_op_reg_opr64_reg64 rRsp) (offset + 0))\n (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 466 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_blocks_auth auth_b abytes_b hkeys_b h_LE) (fun (va_s:va_state)\n (auth_quad_seq:(seq quad32)) -> let (y_0:quad32) = Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 0 0 0 0 in let (y_auth_bytes:quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s) in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 473 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 8))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 474 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRsi) (va_op_reg_opr64_reg64 rRsp) (offset + 16))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 475 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp) (offset + 24))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 476 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRcx) (va_op_opr64_reg64 rR13)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 477 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (fun (va_s:va_state) _ -> let\n (iv_BE:Vale.X64.Decls.quad32) = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2\n va_old_s) in let (ctr_BE_1:quad32) = iv_BE in let (ctr_BE_2:quad32) = Vale.AES.GCTR_s.inc32\n iv_BE 1 in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 483 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 2) (va_op_xmm_xmm 1) (va_op_reg_opr64_reg64\n rR8) 0 Public iv_b 0) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 485 column 20 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Store128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_reg_opr64_reg64 rRbp)\n (va_op_xmm_xmm 1) 0 Secret scratch_b 0) (fun (va_s:va_state) _ -> let (j0:quad32) = va_get_xmm\n 1 va_s in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 487 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load_one_lsb (va_op_xmm_xmm 10)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 489 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_VPaddd (va_op_xmm_xmm 1) (va_op_xmm_xmm 1) (va_op_xmm_xmm 10)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 491 column 26 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_AES_GCM_decrypt_6mult alg h_LE iv_b in128x6_b out128x6_b scratch_b key round_keys\n keys_b hkeys_b) (fun (va_s:va_state) _ -> let (y_cipher128x6:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s) in let (auth_in:(seq quad32)) =\n auth_quad_seq in let (va_arg138:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b in let\n (va_arg137:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = auth_in in let\n (va_arg136:Vale.Def.Types_s.quad32) = y_auth_bytes in let (va_arg135:Vale.Def.Types_s.quad32) =\n y_0 in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 494 column 36 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GHash.lemma_ghash_incremental0_append h_LE va_arg135 va_arg136\n y_cipher128x6 va_arg137 va_arg138) (let auth_in = FStar.Seq.Base.append #quad32 auth_in\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b) in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 498 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 11)\n (va_op_reg_opr64_reg64 rRbp) 32 Secret scratch_b 2) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 499 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR8) (va_op_opr64_reg64 rRcx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 500 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp) (offset + 32))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 501 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 40))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 502 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp) (offset + 48))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 503 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR14) (va_op_opr64_reg64 rRdx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 504 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_InitPshufbMask (va_op_xmm_xmm 9) (va_op_reg_opr64_reg64 rR12)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 505 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Pshufb (va_op_xmm_xmm 11) (va_op_xmm_xmm 9)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 506 column 18 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_blocks128 alg in128_b out128_b key round_keys keys_b hkeys_b h_LE) (fun\n (va_s:va_state) _ -> let (y_cipher128:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s) in let (va_arg134:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b in let\n (va_arg133:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = auth_in in let\n (va_arg132:Vale.Def.Types_s.quad32) = y_0 in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 508 column 36 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GHash.lemma_ghash_incremental0_append h_LE va_arg132 y_cipher128x6\n y_cipher128 va_arg133 va_arg134) (let auth_in = FStar.Seq.Base.append #quad32 auth_in\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b) in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 512 column 22 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Stack_lemma (va_op_reg64_reg64 rRsp) (offset + 24) Public) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 512 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Add64 (va_op_dst_opr64_reg64 rR14) (va_opr_code_Stack (va_op_reg64_reg64 rRsp)\n (offset + 24) Public)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 513 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_IMul64 (va_op_dst_opr64_reg64 rR14) (va_const_opr64 16)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 514 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rR13) (va_op_reg_opr64_reg64 rRsp) (offset + 64))\n (fun (va_s:va_state) _ -> let (y_inout:Vale.Def.Types_s.quad32) = y_cipher128 in let\n (plain_byte_seq:(seq quad32)) = empty_seq_quad32 in let (cipher_byte_seq:(seq quad32)) =\n empty_seq_quad32 in let (va_arg131:Vale.Def.Types_s.quad32) = va_get_xmm 11 va_s in let\n (va_arg130:(FStar.Seq.Base.seq Vale.Def.Types_s.nat32)) = key in let\n (va_arg129:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = cipher_byte_seq in let\n (va_arg128:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = plain_byte_seq in let\n (va_arg127:Vale.AES.AES_common_s.algorithm) = alg in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 519 column 29 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.gctr_partial_opaque_init va_arg127 va_arg128 va_arg129 va_arg130\n va_arg131) (let (total_bytes:(va_int_at_least 0)) = FStar.Seq.Base.length #quad32 auth_quad_seq\n `op_Multiply` 16 + plain_num_bytes in va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 523 column 8 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_qIf va_mods (Cmp_gt (va_op_cmp_reg64 rR13) (va_op_cmp_reg64 rR14)) (qblock va_mods (fun\n (va_s:va_state) -> va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 525 column 21 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp) (offset + 56))\n (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 526 column 14 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR10) (va_op_opr64_reg64 rR13)) (fun (va_s:va_state) _\n -> va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 527 column 26 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.Poly1305.Math.lemma_poly_bits64 ()) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 528 column 14 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_And64 (va_op_dst_opr64_reg64 rR10) (va_const_opr64 15)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 532 column 24 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b total_bytes y_0 auth_in\n h_LE) (fun (va_s:va_state) _ -> let y_inout = Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm\n 8 va_s) in let (raw_auth_quads:(FStar.Seq.Base.seq quad32)) = FStar.Seq.Base.append #quad32\n auth_in (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_old_s) inout_b) in va_qAssertSquash\n va_range1\n \"***** EXPRESSION PRECONDITIONS NOT MET WITHIN line 536 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n ((fun a_1906 (s_1907:(FStar.Seq.Base.seq a_1906)) (i_1908:Prims.nat) (j_1909:Prims.nat) -> let\n (j_1869:Prims.nat) = j_1909 in Prims.b2t (Prims.op_AmpAmp (Prims.op_LessThanOrEqual i_1908\n j_1869) (Prims.op_LessThanOrEqual j_1869 (FStar.Seq.Base.length #a_1906 s_1907))))\n Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_auth_quads) 0 total_bytes)\n (fun _ -> let (auth_input_bytes:(FStar.Seq.Base.seq Vale.Def.Types_s.nat8)) =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n raw_auth_quads) 0 total_bytes in let (padded_auth_bytes:(FStar.Seq.Base.seq\n Vale.Def.Types_s.nat8)) = Vale.AES.GCTR_s.pad_to_128_bits auth_input_bytes in let auth_in =\n Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_auth_bytes in let plain_byte_seq =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_old_s) inout_b in let cipher_byte_seq =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s) inout_b in va_QEmpty ((auth_in,\n cipher_byte_seq, plain_byte_seq, y_inout)))))))))) (qblock va_mods (fun (va_s:va_state) ->\n va_QEmpty ((auth_in, cipher_byte_seq, plain_byte_seq, y_inout))))) (fun (va_s:va_state) va_g ->\n let ((auth_in:(seq quad32)), (cipher_byte_seq:(seq quad32)), (plain_byte_seq:(seq quad32)),\n (y_inout:Vale.Def.Types_s.quad32)) = va_g in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 547 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR11) (va_op_opr64_reg64 rR15)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 548 column 25 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_make_length_quad ()) (fun (va_s:va_state) _ -> let\n (length_quad32:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 0\n va_s) in va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 551 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Ghash_register hkeys_b h_LE y_inout) (fun (va_s:va_state) _ -> let\n (y_final:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s)\n in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 554 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64\n rRbp) 0 Secret scratch_b 0) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 557 column 18 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gctr_register alg key round_keys keys_b) (fun (va_s:va_state) _ -> let\n (va_arg126:Vale.Def.Types_s.quad32) = va_get_xmm 8 va_s in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 560 column 40 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.Arch.Types.le_seq_quad32_to_bytes_of_singleton va_arg126)\n (va_qAssertSquash va_range1\n \"***** EXPRESSION PRECONDITIONS NOT MET WITHIN line 561 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n ((fun (icb_BE_677:Vale.Def.Types_s.quad32) (plain_LE_678:Vale.Def.Types_s.quad32)\n (alg_679:Vale.AES.AES_common_s.algorithm) (key_680:(FStar.Seq.Base.seq Vale.Def.Types_s.nat32))\n (i_681:Prims.int) -> Vale.AES.AES_s.is_aes_key_LE alg_679 key_680) j0 y_final alg key 0) (fun _\n -> va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 561 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_xmm 8 va_s == Vale.AES.GCTR_s.gctr_encrypt_block j0 y_final alg key 0) (let\n (plain128:(FStar.Seq.Base.seq Vale.X64.Decls.quad32)) = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b) in let\n (cipher128:(FStar.Seq.Base.seq Vale.X64.Decls.quad32)) = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) in128_b) in va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 566 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (FStar.Seq.Base.length #quad32 plain_byte_seq == 0 ==> FStar.Seq.Base.equal\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 plain128 plain_byte_seq)\n plain128) (va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 567 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (FStar.Seq.Base.length #quad32 cipher_byte_seq == 0 ==> FStar.Seq.Base.equal\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 cipher128 cipher_byte_seq)\n cipher128) (let (va_arg125:Vale.Def.Types_s.quad32) = Vale.AES.GCTR.inc32lite ctr_BE_2 len128x6\n in let (va_arg124:Vale.Def.Types_s.quad32) = ctr_BE_2 in let (va_arg123:(FStar.Seq.Base.seq\n Vale.Def.Types_s.nat32)) = key in let (va_arg122:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32))\n = Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) out128_b in let\n (va_arg121:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_old_s) in128_b in let (va_arg120:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s) out128x6_b in let\n (va_arg119:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_old_s) in128x6_b in let (va_arg118:Prims.nat) = len128 in let\n (va_arg117:Prims.nat) = len128x6 in let (va_arg116:Vale.AES.AES_common_s.algorithm) = alg in\n va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 569 column 30 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.lemma_gctr_partial_append va_arg116 va_arg117 va_arg118\n va_arg119 va_arg120 va_arg121 va_arg122 va_arg123 va_arg124 va_arg125) (let\n (va_arg115:Vale.Def.Types_s.quad32) = Vale.AES.GCTR.inc32lite (Vale.AES.GCTR.inc32lite ctr_BE_2\n len128x6) len128 in let (va_arg114:Vale.Def.Types_s.quad32) = ctr_BE_2 in let\n (va_arg113:(FStar.Seq.Base.seq Vale.Def.Types_s.nat32)) = key in let\n (va_arg112:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = cipher_byte_seq in let\n (va_arg111:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = plain_byte_seq in let\n (va_arg110:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s) out128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) out128_b) in let\n (va_arg109:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b) in let (va_arg108:Prims.nat) =\n FStar.Seq.Base.length #quad32 plain_byte_seq in let (va_arg107:Prims.nat) = len128x6 + len128\n in let (va_arg106:Vale.AES.AES_common_s.algorithm) = alg in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 575 column 30 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.lemma_gctr_partial_append va_arg106 va_arg107 va_arg108\n va_arg109 va_arg110 va_arg111 va_arg112 va_arg113 va_arg114 va_arg115) (let\n (va_arg105:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = auth_in in let\n (va_arg104:Vale.Def.Types_s.quad32) = y_0 in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 583 column 23 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GHash.lemma_hash_append2 h_LE va_arg104 y_inout y_final va_arg105\n length_quad32) (let auth_in = FStar.Seq.Base.append #quad32 auth_in (FStar.Seq.Base.create\n #Vale.Def.Types_s.quad32 1 length_quad32) in let (va_arg103:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = auth_in in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 585 column 31 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GHash.ghash_incremental_to_ghash h_LE va_arg103) (va_QEmpty\n (()))))))))))))))))))))))))))))))))))))))))))))))))\n\n\nval va_lemma_Gcm_blocks : va_b0:va_code -> va_s0:va_state -> alg:algorithm -> offset:int ->\n auth_b:buffer128 -> abytes_b:buffer128 -> in128x6_b:buffer128 -> out128x6_b:buffer128 ->\n in128_b:buffer128 -> out128_b:buffer128 -> inout_b:buffer128 -> iv_b:buffer128 ->\n scratch_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_blocks alg offset) va_s0 /\\ va_get_ok va_s0 /\\\n (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in let (h_LE:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.X64.Decls.buffer128_read hkeys_b 2\n (va_get_mem_heaplet 0 va_s0)) in sse_enabled /\\ movbe_enabled /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0)\n Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp\n va_s0 + offset + 8) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 24) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 40) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 56) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet\n 1 va_s0) (va_get_reg64 rRdi va_s0) auth_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0)\n Secret /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 7 va_s0) abytes_ptr abytes_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 2\n va_s0) (va_get_reg64 rR8 va_s0) iv_b 1 (va_get_mem_layout va_s0) Public /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 6 va_s0) in128x6_ptr in128x6_b len128x6\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 6\n va_s0) out128x6_ptr out128x6_b len128x6 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0) in128_ptr in128_b len128\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1\n va_s0) out128_ptr out128_b len128 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) inout_ptr inout_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 3\n va_s0) (va_get_reg64 rRbp va_s0) scratch_b 9 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0) hkeys_b 8\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.buffer_disjoints128 iv_b ([keys_b;\n scratch_b; in128x6_b; out128x6_b; hkeys_b; in128_b; out128_b; inout_b]) /\\\n Vale.X64.Decls.buffer_disjoints128 scratch_b ([keys_b; in128x6_b; out128x6_b; in128_b;\n out128_b; inout_b; hkeys_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128x6_b ([keys_b;\n hkeys_b; in128_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128_b ([keys_b; hkeys_b;\n out128x6_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 inout_b ([keys_b; hkeys_b;\n out128x6_b; out128_b]) /\\ (Vale.X64.Decls.buffers_disjoint128 in128x6_b out128x6_b \\/ in128x6_b\n == out128x6_b) /\\ (Vale.X64.Decls.buffers_disjoint128 in128_b out128_b \\/ in128_b == out128_b)\n /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ in128x6_ptr\n + 16 `op_Multiply` len128x6 < pow2_64 /\\ out128x6_ptr + 16 `op_Multiply` len128x6 < pow2_64 /\\\n in128_ptr + 16 `op_Multiply` len128 < pow2_64 /\\ out128_ptr + 16 `op_Multiply` len128 < pow2_64\n /\\ inout_ptr + 16 < pow2_64 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 auth_b ==\n va_get_reg64 rRdx va_s0 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 abytes_b == 1\n /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128x6_b ==\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out128x6_b /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 out128_b /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128\n in128x6_b == len128x6 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b ==\n len128 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 inout_b == 1 /\\\n plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32 /\\ va_get_reg64 rR9 va_s0 + 32 <\n pow2_64 /\\ Vale.X64.Memory.buffer_addr #Vale.X64.Memory.vuint128 keys_b (va_get_mem_heaplet 0\n va_s0) + 128 < pow2_64 /\\ len128x6 `op_Modulus` 6 == 0 /\\ (len128x6 > 0 ==> len128x6 >= 6) /\\\n 12 + len128x6 + 6 < pow2_32 /\\ (va_mul_nat len128x6 (128 `op_Division` 8) + va_mul_nat len128\n (128 `op_Division` 8) <= plain_num_bytes /\\ plain_num_bytes < va_mul_nat len128x6 (128\n `op_Division` 8) + va_mul_nat len128 (128 `op_Division` 8) + 128 `op_Division` 8) /\\\n (va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) <= va_get_reg64 rRsi va_s0 /\\\n va_get_reg64 rRsi va_s0 < va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) + 128\n `op_Division` 8) /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rRcx va_s0)\n (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 (Vale.AES.AES_s.aes_encrypt_LE alg key\n (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0))))))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in let (h_LE:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.X64.Decls.buffer128_read hkeys_b 2\n (va_get_mem_heaplet 0 va_s0)) in Vale.X64.Decls.modifies_buffer128 out128_b (va_get_mem_heaplet\n 1 va_s0) (va_get_mem_heaplet 1 va_sM) /\\ Vale.X64.Decls.modifies_buffer128 iv_b\n (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2 va_sM) /\\ Vale.X64.Decls.modifies_buffer128\n scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3 va_sM) /\\\n Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM) /\\ Vale.X64.Decls.modifies_buffer128 out128x6_b (va_get_mem_heaplet 6 va_s0)\n (va_get_mem_heaplet 6 va_sM) /\\ plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32\n /\\ (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in let\n (ctr_BE_1:quad32) = iv_BE in let (ctr_BE_2:quad32) = Vale.AES.GCTR_s.inc32 iv_BE 1 in let\n (plain_in:(seq quad32)) = (if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) then FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b) else FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in128_b)) in let (cipher_out:(seq quad32)) = (if (plain_num_bytes\n > (len128x6 + len128) `op_Multiply` 128 `op_Division` 8) then FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM)\n out128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM) inout_b) else\n FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM)\n out128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out128_b)) in let\n (cipher_bound:nat) = (if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128 `op_Division`\n 8) then (len128x6 + len128 + 1) else (len128x6 + len128)) in Vale.AES.GCTR.gctr_partial alg\n cipher_bound plain_in cipher_out key ctr_BE_2 /\\ (let (length_quad:quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.Def.Types_s.insert_nat64\n (Vale.Def.Types_s.insert_nat64 (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0) (8\n `op_Multiply` va_get_reg64 rRsi va_s0) 1) (8 `op_Multiply` plain_num_bytes) 0) in let\n (raw_auth_quads:(seq quad32)) = (if (va_get_reg64 rRsi va_s0 > va_get_reg64 rRdx va_s0\n `op_Multiply` 128 `op_Division` 8) then FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 7 va_s0) abytes_b) else Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0)\n auth_b) in let (auth_input_bytes:(seq nat8)) = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_auth_quads) 0 (va_get_reg64 rRsi va_s0) in let\n (padded_auth_bytes:(seq nat8)) = Vale.AES.GCTR_s.pad_to_128_bits auth_input_bytes in let\n (auth_quad_seq:(seq quad32)) = Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_auth_bytes in let\n (raw_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32 (FStar.Seq.Base.append #quad32\n auth_quad_seq (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b))\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b) in let (total_bytes:nat) =\n FStar.Seq.Base.length #quad32 auth_quad_seq `op_Multiply` 16 + plain_num_bytes in let\n (raw_quad_seq:(seq quad32)) = (if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) then (let (ab:(seq nat8)) = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes (FStar.Seq.Base.append #quad32 raw_quad_seq\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b))) 0 total_bytes in let (pb:(seq\n nat8)) = Vale.AES.GCTR_s.pad_to_128_bits ab in Vale.Def.Types_s.le_bytes_to_seq_quad32 pb) else\n raw_quad_seq) in let (auth_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32 raw_quad_seq\n (FStar.Seq.Base.create #quad32 1 length_quad) in va_get_xmm 8 va_sM ==\n Vale.AES.GCTR_s.gctr_encrypt_block ctr_BE_1 (Vale.AES.GHash_s.ghash_LE h_LE auth_quad_seq) alg\n key 0))) /\\ va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 6 va_sM\n (va_update_mem_heaplet 5 va_sM (va_update_mem_heaplet 3 va_sM (va_update_mem_heaplet 2 va_sM\n (va_update_mem_heaplet 1 va_sM (va_update_xmm 15 va_sM (va_update_xmm 14 va_sM (va_update_xmm\n 13 va_sM (va_update_xmm 12 va_sM (va_update_xmm 11 va_sM (va_update_xmm 10 va_sM (va_update_xmm\n 9 va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5\n va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1\n va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR15 va_sM (va_update_reg64 rR14 va_sM\n (va_update_reg64 rR13 va_sM (va_update_reg64 rR12 va_sM (va_update_reg64 rR11 va_sM\n (va_update_reg64 rR10 va_sM (va_update_reg64 rR9 va_sM (va_update_reg64 rR8 va_sM\n (va_update_reg64 rRbp va_sM (va_update_reg64 rRsi va_sM (va_update_reg64 rRdi va_sM\n (va_update_reg64 rRdx va_sM (va_update_reg64 rRcx va_sM (va_update_reg64 rRbx va_sM\n (va_update_reg64 rRax va_sM (va_update_ok va_sM (va_update_mem va_sM\n va_s0)))))))))))))))))))))))))))))))))))))))))\n[@\"opaque_to_smt\"]\nlet va_lemma_Gcm_blocks va_b0 va_s0 alg offset auth_b abytes_b in128x6_b out128x6_b in128_b\n out128_b inout_b iv_b scratch_b key round_keys keys_b hkeys_b =\n let (va_mods:va_mods_t) = [va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5;\n va_Mod_mem_heaplet 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14;\n va_Mod_xmm 13; va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8;\n va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm\n 1; va_Mod_xmm 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12;\n va_Mod_reg64 rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp;\n va_Mod_reg64 rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx;\n va_Mod_reg64 rRax; va_Mod_ok; va_Mod_mem] in\n let va_qc = va_qcode_Gcm_blocks va_mods alg offset auth_b abytes_b in128x6_b out128x6_b in128_b\n out128_b inout_b iv_b scratch_b key round_keys keys_b hkeys_b in\n let (va_sM, va_fM, va_g) = va_wp_sound_code_norm (va_code_Gcm_blocks alg offset) va_qc va_s0 (fun\n va_s0 va_sM va_g -> let () = va_g in label va_range1\n \"***** POSTCONDITION NOT MET AT line 283 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\ (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64\n (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0) in let\n (in128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 8) (va_get_stack va_s0) in let (out128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack va_s0) in\n let (len128x6:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 24) (va_get_stack va_s0) in let (in128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack va_s0) in\n let (out128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 40) (va_get_stack va_s0) in let (len128:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack va_s0) in\n let (inout_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 56) (va_get_stack va_s0) in let (plain_num_bytes:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack va_s0) in\n let (h_LE:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.X64.Decls.buffer128_read hkeys_b 2 (va_get_mem_heaplet 0 va_s0)) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 396 column 56 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out128_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 397 column 52 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 iv_b (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 398 column 57 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 399 column 55 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 400 column 58 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out128x6_b (va_get_mem_heaplet 6 va_s0) (va_get_mem_heaplet\n 6 va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 403 column 39 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (plain_num_bytes < pow2_32) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 404 column 38 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRsi va_s0 < pow2_32) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 406 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 408 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (ctr_BE_1:quad32) = iv_BE in label va_range1\n \"***** POSTCONDITION NOT MET AT line 409 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (ctr_BE_2:quad32) = Vale.AES.GCTR_s.inc32 iv_BE 1 in label va_range1\n \"***** POSTCONDITION NOT MET AT line 412 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (plain_in:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b)) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in128_b)) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 421 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (cipher_out:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM) out128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b)) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_sM) out128_b)) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 430 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (cipher_bound:nat) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> len128x6 + len128 + 1) (fun _ -> len128x6 + len128) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 434 column 77 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.AES.GCTR.gctr_partial alg cipher_bound plain_in cipher_out key ctr_BE_2) /\\ label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 438 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (length_quad:quad32) = Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.Def.Types_s.insert_nat64 (Vale.Def.Types_s.insert_nat64 (Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 0 0 0 0) (8 `op_Multiply` va_get_reg64 rRsi va_s0) 1) (8 `op_Multiply`\n plain_num_bytes) 0) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 440 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (raw_auth_quads:(seq quad32)) = va_if (va_get_reg64 rRsi va_s0 > va_get_reg64 rRdx va_s0\n `op_Multiply` 128 `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 7 va_s0) abytes_b)) (fun _ -> Vale.X64.Decls.s128 (va_get_mem_heaplet 1\n va_s0) auth_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 444 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (auth_input_bytes:(seq nat8)) = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_auth_quads) 0 (va_get_reg64 rRsi va_s0) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 445 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (padded_auth_bytes:(seq nat8)) = Vale.AES.GCTR_s.pad_to_128_bits auth_input_bytes in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 446 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (auth_quad_seq:(seq quad32)) = Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_auth_bytes\n in label va_range1\n \"***** POSTCONDITION NOT MET AT line 448 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (raw_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32 (FStar.Seq.Base.append #quad32\n auth_quad_seq (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b))\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 452 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (total_bytes:nat) = FStar.Seq.Base.length #quad32 auth_quad_seq `op_Multiply` 16 +\n plain_num_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 453 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (raw_quad_seq:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply`\n 128 `op_Division` 8) (fun _ -> let (ab:(seq nat8)) = FStar.Seq.Base.slice\n #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes (FStar.Seq.Base.append #quad32\n raw_quad_seq (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b))) 0 total_bytes in let\n (pb:(seq nat8)) = Vale.AES.GCTR_s.pad_to_128_bits ab in Vale.Def.Types_s.le_bytes_to_seq_quad32\n pb) (fun _ -> raw_quad_seq) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 460 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (auth_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32 raw_quad_seq\n (FStar.Seq.Base.create #quad32 1 length_quad) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 461 column 106 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_xmm 8 va_sM == Vale.AES.GCTR_s.gctr_encrypt_block ctr_BE_1 (Vale.AES.GHash_s.ghash_LE\n h_LE auth_quad_seq) alg key 0)))))))))))))))))) in\n assert_norm (va_qc.mods == va_mods);\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet\n 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13;\n va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm\n 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64\n rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64\n rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64\n rRax; va_Mod_ok; va_Mod_mem]) va_sM va_s0;\n (va_sM, va_fM)\n\n[@ va_qattr]\nlet va_wp_Gcm_blocks (alg:algorithm) (offset:int) (auth_b:buffer128) (abytes_b:buffer128)\n (in128x6_b:buffer128) (out128x6_b:buffer128) (in128_b:buffer128) (out128_b:buffer128)\n (inout_b:buffer128) (iv_b:buffer128) (scratch_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (va_s0:va_state) (va_k:(va_state -> unit ->\n Type0)) : Type0 =\n (va_get_ok va_s0 /\\ (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64\n (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0) in let\n (in128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 8) (va_get_stack va_s0) in let (out128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack va_s0) in\n let (len128x6:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 24) (va_get_stack va_s0) in let (in128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack va_s0) in\n let (out128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 40) (va_get_stack va_s0) in let (len128:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack va_s0) in\n let (inout_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 56) (va_get_stack va_s0) in let (plain_num_bytes:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack va_s0) in\n let (h_LE:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.X64.Decls.buffer128_read hkeys_b 2 (va_get_mem_heaplet 0 va_s0)) in sse_enabled /\\\n movbe_enabled /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 0)\n (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64\n (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 24) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 40) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 56) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet\n 1 va_s0) (va_get_reg64 rRdi va_s0) auth_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0)\n Secret /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 7 va_s0) abytes_ptr abytes_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 2\n va_s0) (va_get_reg64 rR8 va_s0) iv_b 1 (va_get_mem_layout va_s0) Public /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 6 va_s0) in128x6_ptr in128x6_b len128x6\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 6\n va_s0) out128x6_ptr out128x6_b len128x6 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0) in128_ptr in128_b len128\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1\n va_s0) out128_ptr out128_b len128 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) inout_ptr inout_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 3\n va_s0) (va_get_reg64 rRbp va_s0) scratch_b 9 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0) hkeys_b 8\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.buffer_disjoints128 iv_b ([keys_b;\n scratch_b; in128x6_b; out128x6_b; hkeys_b; in128_b; out128_b; inout_b]) /\\\n Vale.X64.Decls.buffer_disjoints128 scratch_b ([keys_b; in128x6_b; out128x6_b; in128_b;\n out128_b; inout_b; hkeys_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128x6_b ([keys_b;\n hkeys_b; in128_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128_b ([keys_b; hkeys_b;\n out128x6_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 inout_b ([keys_b; hkeys_b;\n out128x6_b; out128_b]) /\\ (Vale.X64.Decls.buffers_disjoint128 in128x6_b out128x6_b \\/ in128x6_b\n == out128x6_b) /\\ (Vale.X64.Decls.buffers_disjoint128 in128_b out128_b \\/ in128_b == out128_b)\n /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ in128x6_ptr\n + 16 `op_Multiply` len128x6 < pow2_64 /\\ out128x6_ptr + 16 `op_Multiply` len128x6 < pow2_64 /\\\n in128_ptr + 16 `op_Multiply` len128 < pow2_64 /\\ out128_ptr + 16 `op_Multiply` len128 < pow2_64\n /\\ inout_ptr + 16 < pow2_64 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 auth_b ==\n va_get_reg64 rRdx va_s0 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 abytes_b == 1\n /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128x6_b ==\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out128x6_b /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 out128_b /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128\n in128x6_b == len128x6 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b ==\n len128 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 inout_b == 1 /\\\n plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32 /\\ va_get_reg64 rR9 va_s0 + 32 <\n pow2_64 /\\ Vale.X64.Memory.buffer_addr #Vale.X64.Memory.vuint128 keys_b (va_get_mem_heaplet 0\n va_s0) + 128 < pow2_64 /\\ len128x6 `op_Modulus` 6 == 0 /\\ (len128x6 > 0 ==> len128x6 >= 6) /\\\n 12 + len128x6 + 6 < pow2_32 /\\ (va_mul_nat len128x6 (128 `op_Division` 8) + va_mul_nat len128\n (128 `op_Division` 8) <= plain_num_bytes /\\ plain_num_bytes < va_mul_nat len128x6 (128\n `op_Division` 8) + va_mul_nat len128 (128 `op_Division` 8) + 128 `op_Division` 8) /\\\n (va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) <= va_get_reg64 rRsi va_s0 /\\\n va_get_reg64 rRsi va_s0 < va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) + 128\n `op_Division` 8) /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rRcx va_s0)\n (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 (Vale.AES.AES_s.aes_encrypt_LE alg key\n (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0)))) /\\ (forall (va_x_mem:vale_heap)\n (va_x_rax:nat64) (va_x_rbx:nat64) (va_x_rcx:nat64) (va_x_rdx:nat64) (va_x_rdi:nat64)\n (va_x_rsi:nat64) (va_x_rbp:nat64) (va_x_r8:nat64) (va_x_r9:nat64) (va_x_r10:nat64)\n (va_x_r11:nat64) (va_x_r12:nat64) (va_x_r13:nat64) (va_x_r14:nat64) (va_x_r15:nat64)\n (va_x_xmm0:quad32) (va_x_xmm1:quad32) (va_x_xmm2:quad32) (va_x_xmm3:quad32) (va_x_xmm4:quad32)\n (va_x_xmm5:quad32) (va_x_xmm6:quad32) (va_x_xmm7:quad32) (va_x_xmm8:quad32) (va_x_xmm9:quad32)\n (va_x_xmm10:quad32) (va_x_xmm11:quad32) (va_x_xmm12:quad32) (va_x_xmm13:quad32)\n (va_x_xmm14:quad32) (va_x_xmm15:quad32) (va_x_heap1:vale_heap) (va_x_heap2:vale_heap)\n (va_x_heap3:vale_heap) (va_x_heap5:vale_heap) (va_x_heap6:vale_heap)\n (va_x_efl:Vale.X64.Flags.t) . let va_sM = va_upd_flags va_x_efl (va_upd_mem_heaplet 6\n va_x_heap6 (va_upd_mem_heaplet 5 va_x_heap5 (va_upd_mem_heaplet 3 va_x_heap3\n (va_upd_mem_heaplet 2 va_x_heap2 (va_upd_mem_heaplet 1 va_x_heap1 (va_upd_xmm 15 va_x_xmm15\n (va_upd_xmm 14 va_x_xmm14 (va_upd_xmm 13 va_x_xmm13 (va_upd_xmm 12 va_x_xmm12 (va_upd_xmm 11\n va_x_xmm11 (va_upd_xmm 10 va_x_xmm10 (va_upd_xmm 9 va_x_xmm9 (va_upd_xmm 8 va_x_xmm8\n (va_upd_xmm 7 va_x_xmm7 (va_upd_xmm 6 va_x_xmm6 (va_upd_xmm 5 va_x_xmm5 (va_upd_xmm 4 va_x_xmm4\n (va_upd_xmm 3 va_x_xmm3 (va_upd_xmm 2 va_x_xmm2 (va_upd_xmm 1 va_x_xmm1 (va_upd_xmm 0 va_x_xmm0\n (va_upd_reg64 rR15 va_x_r15 (va_upd_reg64 rR14 va_x_r14 (va_upd_reg64 rR13 va_x_r13\n (va_upd_reg64 rR12 va_x_r12 (va_upd_reg64 rR11 va_x_r11 (va_upd_reg64 rR10 va_x_r10\n (va_upd_reg64 rR9 va_x_r9 (va_upd_reg64 rR8 va_x_r8 (va_upd_reg64 rRbp va_x_rbp (va_upd_reg64\n rRsi va_x_rsi (va_upd_reg64 rRdi va_x_rdi (va_upd_reg64 rRdx va_x_rdx (va_upd_reg64 rRcx\n va_x_rcx (va_upd_reg64 rRbx va_x_rbx (va_upd_reg64 rRax va_x_rax (va_upd_mem va_x_mem\n va_s0))))))))))))))))))))))))))))))))))))) in va_get_ok va_sM /\\ (let\n (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in let (h_LE:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.X64.Decls.buffer128_read hkeys_b 2\n (va_get_mem_heaplet 0 va_s0)) in Vale.X64.Decls.modifies_buffer128 out128_b (va_get_mem_heaplet\n 1 va_s0) (va_get_mem_heaplet 1 va_sM) /\\ Vale.X64.Decls.modifies_buffer128 iv_b\n (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2 va_sM) /\\ Vale.X64.Decls.modifies_buffer128\n scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3 va_sM) /\\\n Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM) /\\ Vale.X64.Decls.modifies_buffer128 out128x6_b (va_get_mem_heaplet 6 va_s0)\n (va_get_mem_heaplet 6 va_sM) /\\ plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32\n /\\ (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in let\n (ctr_BE_1:quad32) = iv_BE in let (ctr_BE_2:quad32) = Vale.AES.GCTR_s.inc32 iv_BE 1 in let\n (plain_in:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b)) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in128_b)) in let (cipher_out:(seq quad32)) = va_if\n (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128 `op_Division` 8) (fun _ ->\n FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_sM) out128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM)\n inout_b)) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM)\n out128_b)) in let (cipher_bound:nat) = va_if (plain_num_bytes > (len128x6 + len128)\n `op_Multiply` 128 `op_Division` 8) (fun _ -> len128x6 + len128 + 1) (fun _ -> len128x6 +\n len128) in Vale.AES.GCTR.gctr_partial alg cipher_bound plain_in cipher_out key ctr_BE_2 /\\ (let\n (length_quad:quad32) = Vale.Def.Types_s.reverse_bytes_quad32 (Vale.Def.Types_s.insert_nat64\n (Vale.Def.Types_s.insert_nat64 (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0) (8\n `op_Multiply` va_get_reg64 rRsi va_s0) 1) (8 `op_Multiply` plain_num_bytes) 0) in let\n (raw_auth_quads:(seq quad32)) = va_if (va_get_reg64 rRsi va_s0 > va_get_reg64 rRdx va_s0\n `op_Multiply` 128 `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 7 va_s0) abytes_b)) (fun _ -> Vale.X64.Decls.s128 (va_get_mem_heaplet 1\n va_s0) auth_b) in let (auth_input_bytes:(seq nat8)) = FStar.Seq.Base.slice\n #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_auth_quads) 0 (va_get_reg64\n rRsi va_s0) in let (padded_auth_bytes:(seq nat8)) = Vale.AES.GCTR_s.pad_to_128_bits\n auth_input_bytes in let (auth_quad_seq:(seq quad32)) = Vale.Def.Types_s.le_bytes_to_seq_quad32\n padded_auth_bytes in let (raw_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32\n (FStar.Seq.Base.append #quad32 auth_quad_seq (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0)\n in128x6_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b) in let (total_bytes:nat)\n = FStar.Seq.Base.length #quad32 auth_quad_seq `op_Multiply` 16 + plain_num_bytes in let\n (raw_quad_seq:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> let (ab:(seq nat8)) = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes (FStar.Seq.Base.append #quad32 raw_quad_seq\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b))) 0 total_bytes in let (pb:(seq\n nat8)) = Vale.AES.GCTR_s.pad_to_128_bits ab in Vale.Def.Types_s.le_bytes_to_seq_quad32 pb) (fun\n _ -> raw_quad_seq) in let (auth_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32\n raw_quad_seq (FStar.Seq.Base.create #quad32 1 length_quad) in va_get_xmm 8 va_sM ==\n Vale.AES.GCTR_s.gctr_encrypt_block ctr_BE_1 (Vale.AES.GHash_s.ghash_LE h_LE auth_quad_seq) alg\n key 0))) ==> va_k va_sM (())))\n\n\nval va_wpProof_Gcm_blocks : alg:algorithm -> offset:int -> auth_b:buffer128 -> abytes_b:buffer128\n -> in128x6_b:buffer128 -> out128x6_b:buffer128 -> in128_b:buffer128 -> out128_b:buffer128 ->\n inout_b:buffer128 -> iv_b:buffer128 -> scratch_b:buffer128 -> key:(seq nat32) -> round_keys:(seq\n quad32) -> keys_b:buffer128 -> hkeys_b:buffer128 -> va_s0:va_state -> va_k:(va_state -> unit ->\n Type0)\n -> Ghost (va_state & va_fuel & unit)\n (requires (va_t_require va_s0 /\\ va_wp_Gcm_blocks alg offset auth_b abytes_b in128x6_b out128x6_b\n in128_b out128_b inout_b iv_b scratch_b key round_keys keys_b hkeys_b va_s0 va_k))\n (ensures (fun (va_sM, va_f0, va_g) -> va_t_ensure (va_code_Gcm_blocks alg offset) ([va_Mod_flags;\n va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet 3; va_Mod_mem_heaplet 2;\n va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13; va_Mod_xmm 12; va_Mod_xmm\n 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5;\n va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR15;\n va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64 rR11; va_Mod_reg64 rR10;\n va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64 rRsi; va_Mod_reg64 rRdi;\n va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64 rRax; va_Mod_mem]) va_s0\n va_k ((va_sM, va_f0, va_g))))\n[@\"opaque_to_smt\"]\nlet va_wpProof_Gcm_blocks alg offset auth_b abytes_b in128x6_b out128x6_b in128_b out128_b inout_b\n iv_b scratch_b key round_keys keys_b hkeys_b va_s0 va_k =\n let (va_sM, va_f0) = va_lemma_Gcm_blocks (va_code_Gcm_blocks alg offset) va_s0 alg offset auth_b\n abytes_b in128x6_b out128x6_b in128_b out128_b inout_b iv_b scratch_b key round_keys keys_b\n hkeys_b in\n va_lemma_upd_update va_sM;\n assert (va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 6 va_sM\n (va_update_mem_heaplet 5 va_sM (va_update_mem_heaplet 3 va_sM (va_update_mem_heaplet 2 va_sM\n (va_update_mem_heaplet 1 va_sM (va_update_xmm 15 va_sM (va_update_xmm 14 va_sM (va_update_xmm\n 13 va_sM (va_update_xmm 12 va_sM (va_update_xmm 11 va_sM (va_update_xmm 10 va_sM (va_update_xmm\n 9 va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5\n va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1\n va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR15 va_sM (va_update_reg64 rR14 va_sM\n (va_update_reg64 rR13 va_sM (va_update_reg64 rR12 va_sM (va_update_reg64 rR11 va_sM\n (va_update_reg64 rR10 va_sM (va_update_reg64 rR9 va_sM (va_update_reg64 rR8 va_sM\n (va_update_reg64 rRbp va_sM (va_update_reg64 rRsi va_sM (va_update_reg64 rRdi va_sM\n (va_update_reg64 rRdx va_sM (va_update_reg64 rRcx va_sM (va_update_reg64 rRbx va_sM\n (va_update_reg64 rRax va_sM (va_update_ok va_sM (va_update_mem va_sM\n va_s0))))))))))))))))))))))))))))))))))))))));\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet\n 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13;\n va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm\n 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64\n rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64\n rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64\n rRax; va_Mod_mem]) va_sM va_s0;\n let va_g = () in\n (va_sM, va_f0, va_g)\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_quick_Gcm_blocks (alg:algorithm) (offset:int) (auth_b:buffer128) (abytes_b:buffer128)\n (in128x6_b:buffer128) (out128x6_b:buffer128) (in128_b:buffer128) (out128_b:buffer128)\n (inout_b:buffer128) (iv_b:buffer128) (scratch_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) : (va_quickCode unit (va_code_Gcm_blocks alg\n offset)) =\n (va_QProc (va_code_Gcm_blocks alg offset) ([va_Mod_flags; va_Mod_mem_heaplet 6;\n va_Mod_mem_heaplet 5; va_Mod_mem_heaplet 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1;\n va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13; va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10;\n va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm\n 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64\n rR13; va_Mod_reg64 rR12; va_Mod_reg64 rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64\n rR8; va_Mod_reg64 rRbp; va_Mod_reg64 rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64\n rRcx; va_Mod_reg64 rRbx; va_Mod_reg64 rRax; va_Mod_mem]) (va_wp_Gcm_blocks alg offset auth_b\n abytes_b in128x6_b out128x6_b in128_b out128_b inout_b iv_b scratch_b key round_keys keys_b\n hkeys_b) (va_wpProof_Gcm_blocks alg offset auth_b abytes_b in128x6_b out128x6_b in128_b\n out128_b inout_b iv_b scratch_b key round_keys keys_b hkeys_b))\n#pop-options\n//--\n//-- Gcm_blocks_wrapped\n#push-options \"--z3rlimit 60\"\n\nval va_code_Gcm_blocks_wrapped : alg:algorithm -> offset:int -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_blocks_wrapped alg offset =\n (va_Block (va_CCons (va_code_Gcm_blocks alg offset) (va_CCons (va_Block (va_CNil ())) (va_CCons\n (va_Block (va_CNil ())) (va_CNil ())))))\n\nval va_codegen_success_Gcm_blocks_wrapped : alg:algorithm -> offset:int -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_blocks_wrapped alg offset =\n (va_pbool_and (va_codegen_success_Gcm_blocks alg offset) (va_ttrue ()))\n\nPartial implementation:\nlet va_qcode_Gcm_blocks_wrapped\n (va_mods: va_mods_t)\n (alg: algorithm)\n (offset: int)\n (auth_b abytes_b in128x6_b out128x6_b in128_b out128_b inout_b iv_b: buffer128)\n (iv: supported_iv_LE)\n (scratch_b: buffer128)\n (key: (seq nat32))\n (round_keys: (seq quad32))\n (keys_b hkeys_b: buffer128)\n (expected_tag: (seq nat8))\n : (va_quickCode unit (va_code_Gcm_blocks_wrapped alg offset)) =\n \n"
},
{
"role": "assistant",
"content": "(qblock va_mods\n (fun (va_s: va_state) ->\n let va_old_s:va_state = va_s in\n let abytes_ptr:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 0) (va_get_stack va_s)\n in\n let in128x6_ptr:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 8) (va_get_stack va_s)\n in\n let out128x6_ptr:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 16) (va_get_stack va_s)\n in\n let len128x6:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 24) (va_get_stack va_s)\n in\n let in128_ptr:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 32) (va_get_stack va_s)\n in\n let out128_ptr:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 40) (va_get_stack va_s)\n in\n let len128:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 48) (va_get_stack va_s)\n in\n let inout_ptr:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 56) (va_get_stack va_s)\n in\n let plain_num_bytes:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 64) (va_get_stack va_s)\n in\n va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 739 column 15 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_blocks alg offset auth_b abytes_b in128x6_b out128x6_b in128_b out128_b\n inout_b iv_b scratch_b key round_keys keys_b hkeys_b)\n (fun (va_s: va_state) _ ->\n va_qAssertSquash va_range1\n \"***** EXPRESSION PRECONDITIONS NOT MET WITHIN line 741 column 37 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (8 `op_Multiply` (va_get_reg64 rRsi va_old_s) >= 0 /\\\n 8 `op_Multiply` (va_get_reg64 rRsi va_old_s) <= 18446744073709551615 /\\\n 8 `op_Multiply` plain_num_bytes >= 0 /\\\n 8 `op_Multiply` plain_num_bytes <= 18446744073709551615)\n (fun _ ->\n let va_arg55:Vale.Def.Types_s.quad32 =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.Def.Types_s.insert_nat64 (Vale.Def.Types_s.insert_nat64\n (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0)\n (8 `op_Multiply` (va_get_reg64 rRsi va_old_s))\n 1)\n (8 `op_Multiply` plain_num_bytes)\n 0)\n in\n let va_arg54:Vale.Def.Types_s.quad32 = va_get_xmm 8 va_s in\n let va_arg53:Vale.Def.Types_s.quad32 =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.X64.Decls.buffer128_read hkeys_b\n 2\n (va_get_mem_heaplet 0 va_s))\n in\n let va_arg52:Vale.Def.Types_s.quad32 =\n Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_old_s)\n in\n let va_arg51:Vale.AES.GCM_s.supported_iv_LE = iv in\n let va_arg50:Prims.nat = va_get_reg64 rRsi va_old_s in\n let va_arg49:Prims.nat = plain_num_bytes in\n let va_arg48:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32) =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s) inout_b\n in\n let va_arg47:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32) =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) out128_b\n in\n let va_arg46:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32) =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s) out128x6_b\n in\n let va_arg45:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32) =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_old_s) inout_b\n in\n let va_arg44:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32) =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b\n in\n let va_arg43:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32) =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b\n in\n let va_arg42:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32) =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 7 va_old_s) abytes_b\n in\n let va_arg41:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32) =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) auth_b\n in\n let va_arg40:(FStar.Seq.Base.seq Vale.Def.Words_s.nat32) = key in\n let va_arg39:Vale.AES.AES_common_s.algorithm = alg in\n va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 741 column 37 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_: unit) ->\n Vale.AES.GCM.gcm_blocks_dec_helper_simplified va_arg39 va_arg40 va_arg41\n va_arg42 va_arg43 va_arg44 va_arg45 va_arg46 va_arg47 va_arg48 va_arg49\n va_arg50 va_arg51 va_arg52 va_arg53 va_arg54 va_arg55)\n (let auth_raw_quads:(FStar.Seq.Base.seq Vale.X64.Decls.quad32) =\n FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) auth_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 7 va_old_s) abytes_b)\n in\n va_qAssertSquash va_range1\n \"***** EXPRESSION PRECONDITIONS NOT MET WITHIN line 751 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n ((fun\n a_1906\n (s_1907: (FStar.Seq.Base.seq a_1906))\n (i_1908: Prims.nat)\n (j_1909: Prims.nat)\n ->\n let j_1869:Prims.nat = j_1909 in\n Prims.b2t (Prims.op_AmpAmp (Prims.op_LessThanOrEqual i_1908 j_1869)\n (Prims.op_LessThanOrEqual j_1869\n (FStar.Seq.Base.length #a_1906 s_1907)))) Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes auth_raw_quads)\n 0\n (va_get_reg64 rRsi va_old_s))\n (fun _ ->\n let auth_bytes:(FStar.Seq.Base.seq Vale.Def.Types_s.nat8) =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes auth_raw_quads)\n 0\n (va_get_reg64 rRsi va_old_s)\n in\n let va_arg38:Vale.Def.Types_s.quad32 =\n Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_old_s)\n in\n let va_arg37:Vale.AES.GCM_s.supported_iv_LE = iv in\n let va_arg36:Prims.nat = plain_num_bytes in\n let va_arg35:(FStar.Seq.Base.seq Vale.Def.Words_s.nat8) =\n expected_tag\n in\n let va_arg34:(FStar.Seq.Base.seq Vale.Def.Words_s.nat8) =\n auth_bytes\n in\n let va_arg33:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32) =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s) inout_b\n in\n let va_arg32:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32) =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) out128_b\n in\n let va_arg31:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32) =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s) out128x6_b\n in\n let va_arg30:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32) =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_old_s) inout_b\n in\n let va_arg29:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32) =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b\n in\n let va_arg28:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32) =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b\n in\n let va_arg27:(FStar.Seq.Base.seq Vale.Def.Words_s.nat32) = key in\n let va_arg26:Vale.AES.AES_common_s.algorithm = alg in\n va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 752 column 37 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_: unit) ->\n Vale.AES.GCM.gcm_blocks_helper_dec_simplified va_arg26 va_arg27\n va_arg28 va_arg29 va_arg30 va_arg31 va_arg32 va_arg33 va_arg34\n va_arg35 va_arg36 va_arg37 va_arg38)\n (va_QEmpty (()))))))))"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n va_mods: Vale.X64.QuickCode.va_mods_t ->\n win: Prims.bool ->\n alg: Vale.AES.AES_common_s.algorithm ->\n auth_b: Vale.X64.Memory.buffer128 ->\n auth_bytes: Vale.X64.Memory.nat64 ->\n auth_num: Vale.X64.Memory.nat64 ->\n keys_b: Vale.X64.Memory.buffer128 ->\n iv_b: Vale.X64.Memory.buffer128 ->\n iv: Vale.AES.GCM_s.supported_iv_LE ->\n hkeys_b: Vale.X64.Memory.buffer128 ->\n abytes_b: Vale.X64.Memory.buffer128 ->\n in128x6_b: Vale.X64.Memory.buffer128 ->\n out128x6_b: Vale.X64.Memory.buffer128 ->\n len128x6_num: Vale.X64.Memory.nat64 ->\n in128_b: Vale.X64.Memory.buffer128 ->\n out128_b: Vale.X64.Memory.buffer128 ->\n len128_num: Vale.X64.Memory.nat64 ->\n inout_b: Vale.X64.Memory.buffer128 ->\n cipher_num: Vale.X64.Memory.nat64 ->\n scratch_b: Vale.X64.Memory.buffer128 ->\n tag_b: Vale.X64.Memory.buffer128 ->\n key: FStar.Seq.Base.seq Vale.X64.Memory.nat32\n -> Vale.X64.QuickCode.va_quickCode Prims.unit\n (Vale.AES.X64.GCMdecryptOpt.va_code_Gcm_blocks_decrypt_stdcall win alg)\nEffect: Prims.Tot\n\nContext:\nmodule Vale.AES.X64.GCMdecryptOpt\nopen Vale.Def.Prop_s\nopen Vale.Def.Opaque_s\nopen FStar.Seq\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Seq_s\nopen Vale.Def.Types_s\nopen Vale.Arch.Types\nopen Vale.Arch.HeapImpl\nopen Vale.AES.AES_s\nopen Vale.AES.GCTR_s\nopen Vale.AES.GCTR\nopen Vale.AES.GCM\nopen Vale.AES.GHash_s\nopen Vale.AES.GHash\nopen Vale.AES.GCM_s\nopen Vale.AES.X64.AES\nopen Vale.AES.GF128_s\nopen Vale.AES.GF128\nopen Vale.Poly1305.Math\nopen Vale.AES.GCM_helpers\nopen Vale.AES.X64.GHash\nopen Vale.AES.X64.GCTR\nopen Vale.X64.Machine_s\nopen Vale.X64.Memory\nopen Vale.X64.Stack_i\nopen Vale.X64.State\nopen Vale.X64.Decls\nopen Vale.X64.InsBasic\nopen Vale.X64.InsMem\nopen Vale.X64.InsVector\nopen Vale.X64.InsStack\nopen Vale.X64.InsAes\nopen Vale.X64.QuickCode\nopen Vale.X64.QuickCodes\nopen Vale.AES.X64.GF128_Mul\nopen Vale.X64.Stack\nopen Vale.X64.CPU_Features_s\nopen Vale.Math.Poly2.Bits_s\nopen Vale.AES.X64.AESopt\nopen Vale.AES.X64.AESGCM\nopen Vale.AES.X64.AESopt2\nopen Vale.Lib.Meta\nopen Vale.AES.X64.GCMencryptOpt\nopen Vale.AES.OptPublic\nopen Vale.Lib.Basic\n#reset-options \"--z3rlimit 20 --max_ifuel 0\"\n//-- Gcm_extra_bytes\n\nval va_code_Gcm_extra_bytes : alg:algorithm -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_extra_bytes alg =\n (va_Block (va_CCons (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0)\n (va_op_reg_opr64_reg64 rRax) 0 Secret) (va_CCons (va_code_Mov128 (va_op_xmm_xmm 10)\n (va_op_xmm_xmm 0)) (va_CCons (va_code_Ghash_extra_bytes ()) (va_CCons (va_code_Mov128\n (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_CCons (va_code_Pshufb (va_op_xmm_xmm 0)\n (va_op_xmm_xmm 9)) (va_CCons (va_code_AESEncryptBlock alg) (va_CCons (va_code_Pxor\n (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (va_CCons (va_code_Store128_buffer\n (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax) (va_op_xmm_xmm 10) 0 Secret)\n (va_CNil ()))))))))))\n\nval va_codegen_success_Gcm_extra_bytes : alg:algorithm -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_extra_bytes alg =\n (va_pbool_and (va_codegen_success_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0)\n (va_op_reg_opr64_reg64 rRax) 0 Secret) (va_pbool_and (va_codegen_success_Mov128 (va_op_xmm_xmm\n 10) (va_op_xmm_xmm 0)) (va_pbool_and (va_codegen_success_Ghash_extra_bytes ()) (va_pbool_and\n (va_codegen_success_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_pbool_and\n (va_codegen_success_Pshufb (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (va_pbool_and\n (va_codegen_success_AESEncryptBlock alg) (va_pbool_and (va_codegen_success_Pxor (va_op_xmm_xmm\n 10) (va_op_xmm_xmm 0)) (va_pbool_and (va_codegen_success_Store128_buffer\n (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax) (va_op_xmm_xmm 10) 0 Secret)\n (va_ttrue ())))))))))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_extra_bytes (va_mods:va_mods_t) (alg:algorithm) (inout_b:buffer128) (key:(seq\n nat32)) (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (total_bytes:nat)\n (old_hash:quad32) (completed_quads:(seq quad32)) (h_LE:quad32) : (va_quickCode unit\n (va_code_Gcm_extra_bytes alg)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in let (len:(va_int_range\n 1 1)) = 1 in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 188 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64\n rRax) 0 Secret inout_b 0) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 189 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov128 (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (fun (va_s:va_state) _ -> let\n (hash_input:quad32) = va_get_xmm 0 va_s in va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 193 column 22 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Ghash_extra_bytes hkeys_b total_bytes old_hash h_LE completed_quads) (fun\n (va_s:va_state) _ -> va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 194 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (FStar.Seq.Base.equal #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s)\n inout_b) (FStar.Seq.Base.create #quad32 1 hash_input)) (let (snap:(FStar.Seq.Base.seq\n Vale.X64.Decls.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s) inout_b in va_QSeq\n va_range1\n \"***** PRECONDITION NOT MET AT line 198 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 199 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Pshufb (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (fun (va_s:va_state) _ -> va_QBind\n va_range1\n \"***** PRECONDITION NOT MET AT line 200 column 20 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_AESEncryptBlock alg (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 11 va_s)) key\n round_keys keys_b) (fun (va_s:va_state) _ -> va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 201 column 26 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.AES_s.aes_encrypt_LE_reveal ()) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 204 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Pxor (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 205 column 20 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Store128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax)\n (va_op_xmm_xmm 10) 0 Secret inout_b 0) (fun (va_s:va_state) _ -> va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 207 column 24 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.gctr_partial_reveal ()) (va_QEmpty (()))))))))))))))\n\n\nval va_lemma_Gcm_extra_bytes : va_b0:va_code -> va_s0:va_state -> alg:algorithm ->\n inout_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128 -> total_bytes:nat -> old_hash:quad32 -> completed_quads:(seq quad32) ->\n h_LE:quad32\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_extra_bytes alg) va_s0 /\\ va_get_ok va_s0 /\\ (let\n (len:(va_int_range 1 1)) = 1 in sse_enabled /\\ Vale.X64.Decls.buffers_disjoint128 keys_b\n inout_b /\\ Vale.X64.Decls.buffers_disjoint128 hkeys_b inout_b /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) (va_get_reg64 rRax va_s0) inout_b\n len (va_get_mem_layout va_s0) Secret /\\ len == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 inout_b /\\ va_get_xmm 9 va_s0 == Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\ aes_reqs alg key round_keys\n keys_b (va_get_reg64 rR8 va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\\n pclmulqdq_enabled /\\ Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0\n va_s0) hkeys_b) (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret /\\ va_get_xmm 8 va_s0 == Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.AES.GHash.ghash_incremental0 h_LE old_hash completed_quads) /\\ FStar.Seq.Base.length\n #quad32 completed_quads == total_bytes `op_Division` 16 /\\ total_bytes < 16 `op_Multiply`\n FStar.Seq.Base.length #quad32 completed_quads + 16 /\\ va_get_reg64 rR10 va_s0 == total_bytes\n `op_Modulus` 16 /\\ total_bytes `op_Modulus` 16 =!= 0 /\\ (0 < total_bytes /\\ total_bytes < 16\n `op_Multiply` Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes) /\\ 16 `op_Multiply`\n (Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes - 1) < total_bytes)))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (let (len:(va_int_range 1 1)) = 1 in Vale.X64.Decls.modifies_buffer128 inout_b\n (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5 va_sM) /\\ Vale.AES.GCTR.gctr_partial alg len\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b) key (va_get_xmm 11 va_s0) /\\ (let raw_quads =\n FStar.Seq.Base.append #quad32 completed_quads (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0)\n inout_b) in let input_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads) 0 total_bytes in let padded_bytes =\n Vale.AES.GCTR_s.pad_to_128_bits input_bytes in let input_quads =\n Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in l_and (FStar.Seq.Base.length\n #Vale.Def.Types_s.quad32 input_quads > 0) (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8\n va_sM) == Vale.AES.GHash.ghash_incremental h_LE old_hash input_quads))) /\\ va_state_eq va_sM\n (va_update_flags va_sM (va_update_mem_heaplet 5 va_sM (va_update_xmm 10 va_sM (va_update_xmm 8\n va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5 va_sM (va_update_xmm 4\n va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1 va_sM (va_update_xmm 0\n va_sM (va_update_reg64 rR11 va_sM (va_update_reg64 rRcx va_sM (va_update_ok va_sM\n (va_update_mem va_sM va_s0))))))))))))))))))\n[@\"opaque_to_smt\"]\nlet va_lemma_Gcm_extra_bytes va_b0 va_s0 alg inout_b key round_keys keys_b hkeys_b total_bytes\n old_hash completed_quads h_LE =\n let (va_mods:va_mods_t) = [va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8;\n va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm\n 1; va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_ok; va_Mod_mem] in\n let va_qc = va_qcode_Gcm_extra_bytes va_mods alg inout_b key round_keys keys_b hkeys_b\n total_bytes old_hash completed_quads h_LE in\n let (va_sM, va_fM, va_g) = va_wp_sound_code_norm (va_code_Gcm_extra_bytes alg) va_qc va_s0 (fun\n va_s0 va_sM va_g -> let () = va_g in label va_range1\n \"***** POSTCONDITION NOT MET AT line 121 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\ (let (len:(va_int_range 1 1)) = 1 in label va_range1\n \"***** POSTCONDITION NOT MET AT line 174 column 55 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 177 column 95 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.AES.GCTR.gctr_partial alg len (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM) inout_b) key (va_get_xmm 11 va_s0)) /\\ label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 180 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let raw_quads = FStar.Seq.Base.append #quad32 completed_quads (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 181 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let input_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads) 0 total_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 182 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let padded_bytes = Vale.AES.GCTR_s.pad_to_128_bits input_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 183 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let input_quads = Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 186 column 59 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (l_and (FStar.Seq.Base.length #Vale.Def.Types_s.quad32 input_quads > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE old_hash input_quads)))))))) in\n assert_norm (va_qc.mods == va_mods);\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm\n 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1;\n va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_ok; va_Mod_mem]) va_sM va_s0;\n (va_sM, va_fM)\n\n[@ va_qattr]\nlet va_wp_Gcm_extra_bytes (alg:algorithm) (inout_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (total_bytes:nat) (old_hash:quad32)\n (completed_quads:(seq quad32)) (h_LE:quad32) (va_s0:va_state) (va_k:(va_state -> unit -> Type0))\n : Type0 =\n (va_get_ok va_s0 /\\ (let (len:(va_int_range 1 1)) = 1 in sse_enabled /\\\n Vale.X64.Decls.buffers_disjoint128 keys_b inout_b /\\ Vale.X64.Decls.buffers_disjoint128 hkeys_b\n inout_b /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) (va_get_reg64 rRax\n va_s0) inout_b len (va_get_mem_layout va_s0) Secret /\\ len == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 inout_b /\\ va_get_xmm 9 va_s0 == Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\ aes_reqs alg key round_keys\n keys_b (va_get_reg64 rR8 va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\\n pclmulqdq_enabled /\\ Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0\n va_s0) hkeys_b) (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret /\\ va_get_xmm 8 va_s0 == Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.AES.GHash.ghash_incremental0 h_LE old_hash completed_quads) /\\ FStar.Seq.Base.length\n #quad32 completed_quads == total_bytes `op_Division` 16 /\\ total_bytes < 16 `op_Multiply`\n FStar.Seq.Base.length #quad32 completed_quads + 16 /\\ va_get_reg64 rR10 va_s0 == total_bytes\n `op_Modulus` 16 /\\ total_bytes `op_Modulus` 16 =!= 0 /\\ (0 < total_bytes /\\ total_bytes < 16\n `op_Multiply` Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes) /\\ 16 `op_Multiply`\n (Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes - 1) < total_bytes) /\\ (forall\n (va_x_mem:vale_heap) (va_x_rcx:nat64) (va_x_r11:nat64) (va_x_xmm0:quad32) (va_x_xmm1:quad32)\n (va_x_xmm2:quad32) (va_x_xmm3:quad32) (va_x_xmm4:quad32) (va_x_xmm5:quad32) (va_x_xmm6:quad32)\n (va_x_xmm7:quad32) (va_x_xmm8:quad32) (va_x_xmm10:quad32) (va_x_heap5:vale_heap)\n (va_x_efl:Vale.X64.Flags.t) . let va_sM = va_upd_flags va_x_efl (va_upd_mem_heaplet 5\n va_x_heap5 (va_upd_xmm 10 va_x_xmm10 (va_upd_xmm 8 va_x_xmm8 (va_upd_xmm 7 va_x_xmm7\n (va_upd_xmm 6 va_x_xmm6 (va_upd_xmm 5 va_x_xmm5 (va_upd_xmm 4 va_x_xmm4 (va_upd_xmm 3 va_x_xmm3\n (va_upd_xmm 2 va_x_xmm2 (va_upd_xmm 1 va_x_xmm1 (va_upd_xmm 0 va_x_xmm0 (va_upd_reg64 rR11\n va_x_r11 (va_upd_reg64 rRcx va_x_rcx (va_upd_mem va_x_mem va_s0)))))))))))))) in va_get_ok\n va_sM /\\ (let (len:(va_int_range 1 1)) = 1 in Vale.X64.Decls.modifies_buffer128 inout_b\n (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5 va_sM) /\\ Vale.AES.GCTR.gctr_partial alg len\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b) key (va_get_xmm 11 va_s0) /\\ (let raw_quads =\n FStar.Seq.Base.append #quad32 completed_quads (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0)\n inout_b) in let input_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads) 0 total_bytes in let padded_bytes =\n Vale.AES.GCTR_s.pad_to_128_bits input_bytes in let input_quads =\n Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in l_and (FStar.Seq.Base.length\n #Vale.Def.Types_s.quad32 input_quads > 0) (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8\n va_sM) == Vale.AES.GHash.ghash_incremental h_LE old_hash input_quads))) ==> va_k va_sM (())))\n\n\nval va_wpProof_Gcm_extra_bytes : alg:algorithm -> inout_b:buffer128 -> key:(seq nat32) ->\n round_keys:(seq quad32) -> keys_b:buffer128 -> hkeys_b:buffer128 -> total_bytes:nat ->\n old_hash:quad32 -> completed_quads:(seq quad32) -> h_LE:quad32 -> va_s0:va_state ->\n va_k:(va_state -> unit -> Type0)\n -> Ghost (va_state & va_fuel & unit)\n (requires (va_t_require va_s0 /\\ va_wp_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b\n total_bytes old_hash completed_quads h_LE va_s0 va_k))\n (ensures (fun (va_sM, va_f0, va_g) -> va_t_ensure (va_code_Gcm_extra_bytes alg) ([va_Mod_flags;\n va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5;\n va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR11;\n va_Mod_reg64 rRcx; va_Mod_mem]) va_s0 va_k ((va_sM, va_f0, va_g))))\n[@\"opaque_to_smt\"]\nlet va_wpProof_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b total_bytes old_hash\n completed_quads h_LE va_s0 va_k =\n let (va_sM, va_f0) = va_lemma_Gcm_extra_bytes (va_code_Gcm_extra_bytes alg) va_s0 alg inout_b key\n round_keys keys_b hkeys_b total_bytes old_hash completed_quads h_LE in\n va_lemma_upd_update va_sM;\n assert (va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 5 va_sM (va_update_xmm 10\n va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5\n va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1\n va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR11 va_sM (va_update_reg64 rRcx va_sM\n (va_update_ok va_sM (va_update_mem va_sM va_s0)))))))))))))))));\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm\n 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1;\n va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_mem]) va_sM va_s0;\n let va_g = () in\n (va_sM, va_f0, va_g)\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_quick_Gcm_extra_bytes (alg:algorithm) (inout_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (total_bytes:nat) (old_hash:quad32)\n (completed_quads:(seq quad32)) (h_LE:quad32) : (va_quickCode unit (va_code_Gcm_extra_bytes alg)) =\n (va_QProc (va_code_Gcm_extra_bytes alg) ([va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10;\n va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm\n 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_mem])\n (va_wp_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b total_bytes old_hash\n completed_quads h_LE) (va_wpProof_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b\n total_bytes old_hash completed_quads h_LE))\n//--\n//-- Gcm_blocks128\n\nval va_code_Gcm_blocks128 : alg:algorithm -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_blocks128 alg =\n (va_Block (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRbx) (va_op_opr64_reg64 rRdi))\n (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR12) (va_op_opr64_reg64 rRdx)) (va_CCons\n (va_code_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRax)) (va_CCons\n (va_code_Ghash_buffer ()) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRdi)\n (va_op_opr64_reg64 rRbx)) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRdx)\n (va_op_opr64_reg64 rR12)) (va_CCons (va_code_Gctr_blocks128 alg) (va_CNil ())))))))))\n\nval va_codegen_success_Gcm_blocks128 : alg:algorithm -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_blocks128 alg =\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rRbx) (va_op_opr64_reg64 rRdi))\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR12) (va_op_opr64_reg64 rRdx))\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRax))\n (va_pbool_and (va_codegen_success_Ghash_buffer ()) (va_pbool_and (va_codegen_success_Mov64\n (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRbx)) (va_pbool_and (va_codegen_success_Mov64\n (va_op_dst_opr64_reg64 rRdx) (va_op_opr64_reg64 rR12)) (va_pbool_and\n (va_codegen_success_Gctr_blocks128 alg) (va_ttrue ()))))))))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_blocks128 (va_mods:va_mods_t) (alg:algorithm) (in_b:buffer128) (out_b:buffer128)\n (key:(seq nat32)) (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (h_LE:quad32)\n : (va_quickCode unit (va_code_Gcm_blocks128 alg)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 274 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRbx) (va_op_opr64_reg64 rRdi)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 275 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR12) (va_op_opr64_reg64 rRdx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 276 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRax)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 277 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Ghash_buffer hkeys_b in_b h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8\n va_old_s))) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 278 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRbx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 279 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRdx) (va_op_opr64_reg64 rR12)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 280 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gctr_blocks128 alg in_b out_b key round_keys keys_b) (va_QEmpty (()))))))))))\n\n\nval va_lemma_Gcm_blocks128 : va_b0:va_code -> va_s0:va_state -> alg:algorithm -> in_b:buffer128 ->\n out_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128 -> h_LE:quad32\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_blocks128 alg) va_s0 /\\ va_get_ok va_s0 /\\\n (sse_enabled /\\ Vale.X64.Decls.buffers_disjoint128 keys_b out_b /\\\n Vale.X64.Decls.buffers_disjoint128 hkeys_b out_b /\\ (Vale.X64.Decls.buffers_disjoint128 in_b\n out_b \\/ in_b == out_b) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0)\n (va_get_reg64 rRax va_s0) in_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1 va_s0) (va_get_reg64 rRdi va_s0) out_b\n (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\ va_get_reg64 rRax va_s0 + 16\n `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply`\n va_get_reg64 rRdx va_s0 < pow2_64 /\\ l_and (Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 in_b == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out_b)\n (Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b < pow2_32) /\\ va_get_reg64 rRdx\n va_s0 == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b /\\ va_get_xmm 9 va_s0 ==\n Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\\n va_get_reg64 rRdx va_s0 < pow2_32 /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rR8\n va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret)))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (Vale.X64.Decls.modifies_buffer128 out_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM) /\\ Vale.AES.GCTR.gctr_partial alg (va_get_reg64 rRdx va_s0) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b) key\n (va_get_xmm 11 va_s0) /\\ va_get_xmm 11 va_sM == Vale.AES.GCTR.inc32lite (va_get_xmm 11 va_s0)\n (va_get_reg64 rRdx va_s0) /\\ (va_get_reg64 rRdx va_s0 == 0 ==> l_and (va_get_xmm 8 va_sM ==\n va_get_xmm 8 va_s0) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b ==\n Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) out_b)) /\\ (va_get_reg64 rRdx va_s0 > 0 ==>\n l_and (va_get_reg64 rRdx va_s0 <= FStar.Seq.Base.length #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) ==> FStar.Seq.Base.length\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.slice #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) 0 (va_get_reg64 rRdx va_s0)) > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s0)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b)))) /\\ va_state_eq va_sM (va_update_flags va_sM\n (va_update_mem_heaplet 1 va_sM (va_update_xmm 10 va_sM (va_update_xmm 11 va_sM (va_update_xmm 8\n va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5 va_sM (va_update_xmm 4\n va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1 va_sM (va_update_xmm 0\n va_sM (va_update_reg64 rR12 va_sM (va_update_reg64 rRdx va_sM (va_update_reg64 rR10 va_sM\n (va_update_reg64 rR11 va_sM (va_update_reg64 rRdi va_sM (va_update_reg64 rRbx va_sM\n (va_update_ok va_sM (va_update_mem va_sM va_s0)))))))))))))))))))))))\n[@\"opaque_to_smt\"]\nlet va_lemma_Gcm_blocks128 va_b0 va_s0 alg in_b out_b key round_keys keys_b hkeys_b h_LE =\n let (va_mods:va_mods_t) = [va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11;\n va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm\n 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10;\n va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_ok; va_Mod_mem] in\n let va_qc = va_qcode_Gcm_blocks128 va_mods alg in_b out_b key round_keys keys_b hkeys_b h_LE in\n let (va_sM, va_fM, va_g) = va_wp_sound_code_norm (va_code_Gcm_blocks128 alg) va_qc va_s0 (fun\n va_s0 va_sM va_g -> let () = va_g in label va_range1\n \"***** POSTCONDITION NOT MET AT line 210 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\ (label va_range1\n \"***** POSTCONDITION NOT MET AT line 255 column 53 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 261 column 95 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.AES.GCTR.gctr_partial alg (va_get_reg64 rRdx va_s0) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b) key\n (va_get_xmm 11 va_s0)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 262 column 45 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_xmm 11 va_sM == Vale.AES.GCTR.inc32lite (va_get_xmm 11 va_s0) (va_get_reg64 rRdx\n va_s0)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 265 column 93 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRdx va_s0 == 0 ==> l_and (va_get_xmm 8 va_sM == va_get_xmm 8 va_s0)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b == Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) out_b)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 267 column 131 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRdx va_s0 > 0 ==> l_and (va_get_reg64 rRdx va_s0 <= FStar.Seq.Base.length\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) ==>\n FStar.Seq.Base.length #Vale.X64.Decls.quad32 (FStar.Seq.Base.slice #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) 0 (va_get_reg64 rRdx va_s0)) > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s0)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b))))) in\n assert_norm (va_qc.mods == va_mods);\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11; va_Mod_xmm\n 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2;\n va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10;\n va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_ok; va_Mod_mem]) va_sM va_s0;\n (va_sM, va_fM)\n\n[@ va_qattr]\nlet va_wp_Gcm_blocks128 (alg:algorithm) (in_b:buffer128) (out_b:buffer128) (key:(seq nat32))\n (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (h_LE:quad32) (va_s0:va_state)\n (va_k:(va_state -> unit -> Type0)) : Type0 =\n (va_get_ok va_s0 /\\ (sse_enabled /\\ Vale.X64.Decls.buffers_disjoint128 keys_b out_b /\\\n Vale.X64.Decls.buffers_disjoint128 hkeys_b out_b /\\ (Vale.X64.Decls.buffers_disjoint128 in_b\n out_b \\/ in_b == out_b) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0)\n (va_get_reg64 rRax va_s0) in_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1 va_s0) (va_get_reg64 rRdi va_s0) out_b\n (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\ va_get_reg64 rRax va_s0 + 16\n `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply`\n va_get_reg64 rRdx va_s0 < pow2_64 /\\ l_and (Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 in_b == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out_b)\n (Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b < pow2_32) /\\ va_get_reg64 rRdx\n va_s0 == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b /\\ va_get_xmm 9 va_s0 ==\n Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\\n va_get_reg64 rRdx va_s0 < pow2_32 /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rR8\n va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret) /\\ (forall (va_x_mem:vale_heap) (va_x_rbx:nat64) (va_x_rdi:nat64) (va_x_r11:nat64)\n (va_x_r10:nat64) (va_x_rdx:nat64) (va_x_r12:nat64) (va_x_xmm0:quad32) (va_x_xmm1:quad32)\n (va_x_xmm2:quad32) (va_x_xmm3:quad32) (va_x_xmm4:quad32) (va_x_xmm5:quad32) (va_x_xmm6:quad32)\n (va_x_xmm7:quad32) (va_x_xmm8:quad32) (va_x_xmm11:quad32) (va_x_xmm10:quad32)\n (va_x_heap1:vale_heap) (va_x_efl:Vale.X64.Flags.t) . let va_sM = va_upd_flags va_x_efl\n (va_upd_mem_heaplet 1 va_x_heap1 (va_upd_xmm 10 va_x_xmm10 (va_upd_xmm 11 va_x_xmm11\n (va_upd_xmm 8 va_x_xmm8 (va_upd_xmm 7 va_x_xmm7 (va_upd_xmm 6 va_x_xmm6 (va_upd_xmm 5 va_x_xmm5\n (va_upd_xmm 4 va_x_xmm4 (va_upd_xmm 3 va_x_xmm3 (va_upd_xmm 2 va_x_xmm2 (va_upd_xmm 1 va_x_xmm1\n (va_upd_xmm 0 va_x_xmm0 (va_upd_reg64 rR12 va_x_r12 (va_upd_reg64 rRdx va_x_rdx (va_upd_reg64\n rR10 va_x_r10 (va_upd_reg64 rR11 va_x_r11 (va_upd_reg64 rRdi va_x_rdi (va_upd_reg64 rRbx\n va_x_rbx (va_upd_mem va_x_mem va_s0))))))))))))))))))) in va_get_ok va_sM /\\\n (Vale.X64.Decls.modifies_buffer128 out_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM) /\\ Vale.AES.GCTR.gctr_partial alg (va_get_reg64 rRdx va_s0) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b) key\n (va_get_xmm 11 va_s0) /\\ va_get_xmm 11 va_sM == Vale.AES.GCTR.inc32lite (va_get_xmm 11 va_s0)\n (va_get_reg64 rRdx va_s0) /\\ (va_get_reg64 rRdx va_s0 == 0 ==> l_and (va_get_xmm 8 va_sM ==\n va_get_xmm 8 va_s0) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b ==\n Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) out_b)) /\\ (va_get_reg64 rRdx va_s0 > 0 ==>\n l_and (va_get_reg64 rRdx va_s0 <= FStar.Seq.Base.length #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) ==> FStar.Seq.Base.length\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.slice #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) 0 (va_get_reg64 rRdx va_s0)) > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s0)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b)))) ==> va_k va_sM (())))\n\n\nval va_wpProof_Gcm_blocks128 : alg:algorithm -> in_b:buffer128 -> out_b:buffer128 -> key:(seq\n nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 -> hkeys_b:buffer128 -> h_LE:quad32 ->\n va_s0:va_state -> va_k:(va_state -> unit -> Type0)\n -> Ghost (va_state & va_fuel & unit)\n (requires (va_t_require va_s0 /\\ va_wp_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b\n h_LE va_s0 va_k))\n (ensures (fun (va_sM, va_f0, va_g) -> va_t_ensure (va_code_Gcm_blocks128 alg) ([va_Mod_flags;\n va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6;\n va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0;\n va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10; va_Mod_reg64 rR11; va_Mod_reg64 rRdi;\n va_Mod_reg64 rRbx; va_Mod_mem]) va_s0 va_k ((va_sM, va_f0, va_g))))\n[@\"opaque_to_smt\"]\nlet va_wpProof_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b h_LE va_s0 va_k =\n let (va_sM, va_f0) = va_lemma_Gcm_blocks128 (va_code_Gcm_blocks128 alg) va_s0 alg in_b out_b key\n round_keys keys_b hkeys_b h_LE in\n va_lemma_upd_update va_sM;\n assert (va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 1 va_sM (va_update_xmm 10\n va_sM (va_update_xmm 11 va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6\n va_sM (va_update_xmm 5 va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2\n va_sM (va_update_xmm 1 va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR12 va_sM\n (va_update_reg64 rRdx va_sM (va_update_reg64 rR10 va_sM (va_update_reg64 rR11 va_sM\n (va_update_reg64 rRdi va_sM (va_update_reg64 rRbx va_sM (va_update_ok va_sM (va_update_mem\n va_sM va_s0))))))))))))))))))))));\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11; va_Mod_xmm\n 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2;\n va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10;\n va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_mem]) va_sM va_s0;\n let va_g = () in\n (va_sM, va_f0, va_g)\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_quick_Gcm_blocks128 (alg:algorithm) (in_b:buffer128) (out_b:buffer128) (key:(seq nat32))\n (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (h_LE:quad32) : (va_quickCode\n unit (va_code_Gcm_blocks128 alg)) =\n (va_QProc (va_code_Gcm_blocks128 alg) ([va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10;\n va_Mod_xmm 11; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm\n 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64\n rR10; va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_mem])\n (va_wp_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b h_LE)\n (va_wpProof_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b h_LE))\n//--\n//-- Gcm_blocks\n#push-options \"--z3rlimit 1000\"\n\nval va_code_Gcm_blocks : alg:algorithm -> offset:int -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_blocks alg offset =\n (va_Block (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR13) (va_op_opr64_reg64 rRcx))\n (va_CCons (va_code_AddLea64 (va_op_dst_opr64_reg64 rR9) (va_op_opr64_reg64 rR9) (va_const_opr64\n 32)) (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rRbx) (va_op_reg_opr64_reg64 rRsp)\n (offset + 0)) (va_CCons (va_code_Gcm_blocks_auth ()) (va_CCons (va_code_Load64_stack\n (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 8)) (va_CCons\n (va_code_Load64_stack (va_op_dst_opr64_reg64 rRsi) (va_op_reg_opr64_reg64 rRsp) (offset + 16))\n (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp)\n (offset + 24)) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRcx) (va_op_opr64_reg64 rR13))\n (va_CCons (va_code_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (va_CCons\n (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 2) (va_op_xmm_xmm 1) (va_op_reg_opr64_reg64\n rR8) 0 Public) (va_CCons (va_code_Store128_buffer (va_op_heaplet_mem_heaplet 3)\n (va_op_reg_opr64_reg64 rRbp) (va_op_xmm_xmm 1) 0 Secret) (va_CCons (va_code_Load_one_lsb\n (va_op_xmm_xmm 10)) (va_CCons (va_code_VPaddd (va_op_xmm_xmm 1) (va_op_xmm_xmm 1)\n (va_op_xmm_xmm 10)) (va_CCons (va_code_AES_GCM_decrypt_6mult alg) (va_CCons\n (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 11) (va_op_reg_opr64_reg64\n rRbp) 32 Secret) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR8) (va_op_opr64_reg64 rRcx))\n (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp)\n (offset + 32)) (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rRdi)\n (va_op_reg_opr64_reg64 rRsp) (offset + 40)) (va_CCons (va_code_Load64_stack\n (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp) (offset + 48)) (va_CCons\n (va_code_Mov64 (va_op_dst_opr64_reg64 rR14) (va_op_opr64_reg64 rRdx)) (va_CCons\n (va_code_InitPshufbMask (va_op_xmm_xmm 9) (va_op_reg_opr64_reg64 rR12)) (va_CCons\n (va_code_Pshufb (va_op_xmm_xmm 11) (va_op_xmm_xmm 9)) (va_CCons (va_code_Gcm_blocks128 alg)\n (va_CCons (va_code_Stack_lemma ()) (va_CCons (va_code_Add64 (va_op_dst_opr64_reg64 rR14)\n (va_opr_code_Stack (va_op_reg64_reg64 rRsp) (offset + 24) Public)) (va_CCons (va_code_IMul64\n (va_op_dst_opr64_reg64 rR14) (va_const_opr64 16)) (va_CCons (va_code_Load64_stack\n (va_op_dst_opr64_reg64 rR13) (va_op_reg_opr64_reg64 rRsp) (offset + 64)) (va_CCons (va_IfElse\n (va_cmp_gt (va_op_cmp_reg64 rR13) (va_op_cmp_reg64 rR14)) (va_Block (va_CCons\n (va_code_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp) (offset + 56))\n (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR10) (va_op_opr64_reg64 rR13)) (va_CCons\n (va_code_And64 (va_op_dst_opr64_reg64 rR10) (va_const_opr64 15)) (va_CCons\n (va_code_Gcm_extra_bytes alg) (va_CCons (va_Block (va_CNil ())) (va_CNil ()))))))) (va_Block\n (va_CNil ()))) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR11) (va_op_opr64_reg64 rR15))\n (va_CCons (va_code_Gcm_make_length_quad ()) (va_CCons (va_code_Ghash_register ()) (va_CCons\n (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64\n rRbp) 0 Secret) (va_CCons (va_code_Gctr_register alg) (va_CCons (va_Block (va_CNil ()))\n (va_CNil ()))))))))))))))))))))))))))))))))))))\n\nval va_codegen_success_Gcm_blocks : alg:algorithm -> offset:int -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_blocks alg offset =\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR13) (va_op_opr64_reg64 rRcx))\n (va_pbool_and (va_codegen_success_AddLea64 (va_op_dst_opr64_reg64 rR9) (va_op_opr64_reg64 rR9)\n (va_const_opr64 32)) (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64\n rRbx) (va_op_reg_opr64_reg64 rRsp) (offset + 0)) (va_pbool_and\n (va_codegen_success_Gcm_blocks_auth ()) (va_pbool_and (va_codegen_success_Load64_stack\n (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 8)) (va_pbool_and\n (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRsi) (va_op_reg_opr64_reg64 rRsp)\n (offset + 16)) (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRdx)\n (va_op_reg_opr64_reg64 rRsp) (offset + 24)) (va_pbool_and (va_codegen_success_Mov64\n (va_op_dst_opr64_reg64 rRcx) (va_op_opr64_reg64 rR13)) (va_pbool_and (va_codegen_success_Mov128\n (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (va_pbool_and (va_codegen_success_Load128_buffer\n (va_op_heaplet_mem_heaplet 2) (va_op_xmm_xmm 1) (va_op_reg_opr64_reg64 rR8) 0 Public)\n (va_pbool_and (va_codegen_success_Store128_buffer (va_op_heaplet_mem_heaplet 3)\n (va_op_reg_opr64_reg64 rRbp) (va_op_xmm_xmm 1) 0 Secret) (va_pbool_and\n (va_codegen_success_Load_one_lsb (va_op_xmm_xmm 10)) (va_pbool_and (va_codegen_success_VPaddd\n (va_op_xmm_xmm 1) (va_op_xmm_xmm 1) (va_op_xmm_xmm 10)) (va_pbool_and\n (va_codegen_success_AES_GCM_decrypt_6mult alg) (va_pbool_and (va_codegen_success_Load128_buffer\n (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 11) (va_op_reg_opr64_reg64 rRbp) 32 Secret)\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR8) (va_op_opr64_reg64 rRcx))\n (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRax)\n (va_op_reg_opr64_reg64 rRsp) (offset + 32)) (va_pbool_and (va_codegen_success_Load64_stack\n (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 40)) (va_pbool_and\n (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp)\n (offset + 48)) (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR14)\n (va_op_opr64_reg64 rRdx)) (va_pbool_and (va_codegen_success_InitPshufbMask (va_op_xmm_xmm 9)\n (va_op_reg_opr64_reg64 rR12)) (va_pbool_and (va_codegen_success_Pshufb (va_op_xmm_xmm 11)\n (va_op_xmm_xmm 9)) (va_pbool_and (va_codegen_success_Gcm_blocks128 alg) (va_pbool_and\n (va_codegen_success_Stack_lemma ()) (va_pbool_and (va_codegen_success_Add64\n (va_op_dst_opr64_reg64 rR14) (va_opr_code_Stack (va_op_reg64_reg64 rRsp) (offset + 24) Public))\n (va_pbool_and (va_codegen_success_IMul64 (va_op_dst_opr64_reg64 rR14) (va_const_opr64 16))\n (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rR13)\n (va_op_reg_opr64_reg64 rRsp) (offset + 64)) (va_pbool_and (va_pbool_and\n (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp)\n (offset + 56)) (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR10)\n (va_op_opr64_reg64 rR13)) (va_pbool_and (va_codegen_success_And64 (va_op_dst_opr64_reg64 rR10)\n (va_const_opr64 15)) (va_codegen_success_Gcm_extra_bytes alg)))) (va_pbool_and\n (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR11) (va_op_opr64_reg64 rR15)) (va_pbool_and\n (va_codegen_success_Gcm_make_length_quad ()) (va_pbool_and (va_codegen_success_Ghash_register\n ()) (va_pbool_and (va_codegen_success_Load128_buffer (va_op_heaplet_mem_heaplet 3)\n (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64 rRbp) 0 Secret) (va_pbool_and\n (va_codegen_success_Gctr_register alg) (va_ttrue ()))))))))))))))))))))))))))))))))))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_blocks (va_mods:va_mods_t) (alg:algorithm) (offset:int) (auth_b:buffer128)\n (abytes_b:buffer128) (in128x6_b:buffer128) (out128x6_b:buffer128) (in128_b:buffer128)\n (out128_b:buffer128) (inout_b:buffer128) (iv_b:buffer128) (scratch_b:buffer128) (key:(seq nat32))\n (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) : (va_quickCode unit\n (va_code_Gcm_blocks alg offset)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in let\n (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 0) (va_get_stack va_s) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 8) (va_get_stack va_s) in let\n (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 16) (va_get_stack va_s) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 24) (va_get_stack va_s) in let\n (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 32) (va_get_stack va_s) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 40) (va_get_stack va_s) in let\n (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset\n + 48) (va_get_stack va_s) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 56) (va_get_stack va_s) in let\n (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s\n + offset + 64) (va_get_stack va_s) in let (h_LE:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.X64.Decls.buffer128_read hkeys_b 2\n (va_get_mem_heaplet 0 va_old_s)) in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 463 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR13) (va_op_opr64_reg64 rRcx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 464 column 13 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_AddLea64 (va_op_dst_opr64_reg64 rR9) (va_op_opr64_reg64 rR9) (va_const_opr64 32))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 465 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRbx) (va_op_reg_opr64_reg64 rRsp) (offset + 0))\n (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 466 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_blocks_auth auth_b abytes_b hkeys_b h_LE) (fun (va_s:va_state)\n (auth_quad_seq:(seq quad32)) -> let (y_0:quad32) = Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 0 0 0 0 in let (y_auth_bytes:quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s) in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 473 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 8))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 474 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRsi) (va_op_reg_opr64_reg64 rRsp) (offset + 16))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 475 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp) (offset + 24))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 476 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRcx) (va_op_opr64_reg64 rR13)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 477 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (fun (va_s:va_state) _ -> let\n (iv_BE:Vale.X64.Decls.quad32) = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2\n va_old_s) in let (ctr_BE_1:quad32) = iv_BE in let (ctr_BE_2:quad32) = Vale.AES.GCTR_s.inc32\n iv_BE 1 in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 483 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 2) (va_op_xmm_xmm 1) (va_op_reg_opr64_reg64\n rR8) 0 Public iv_b 0) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 485 column 20 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Store128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_reg_opr64_reg64 rRbp)\n (va_op_xmm_xmm 1) 0 Secret scratch_b 0) (fun (va_s:va_state) _ -> let (j0:quad32) = va_get_xmm\n 1 va_s in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 487 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load_one_lsb (va_op_xmm_xmm 10)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 489 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_VPaddd (va_op_xmm_xmm 1) (va_op_xmm_xmm 1) (va_op_xmm_xmm 10)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 491 column 26 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_AES_GCM_decrypt_6mult alg h_LE iv_b in128x6_b out128x6_b scratch_b key round_keys\n keys_b hkeys_b) (fun (va_s:va_state) _ -> let (y_cipher128x6:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s) in let (auth_in:(seq quad32)) =\n auth_quad_seq in let (va_arg138:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b in let\n (va_arg137:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = auth_in in let\n (va_arg136:Vale.Def.Types_s.quad32) = y_auth_bytes in let (va_arg135:Vale.Def.Types_s.quad32) =\n y_0 in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 494 column 36 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GHash.lemma_ghash_incremental0_append h_LE va_arg135 va_arg136\n y_cipher128x6 va_arg137 va_arg138) (let auth_in = FStar.Seq.Base.append #quad32 auth_in\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b) in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 498 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 11)\n (va_op_reg_opr64_reg64 rRbp) 32 Secret scratch_b 2) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 499 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR8) (va_op_opr64_reg64 rRcx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 500 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp) (offset + 32))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 501 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 40))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 502 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp) (offset + 48))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 503 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR14) (va_op_opr64_reg64 rRdx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 504 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_InitPshufbMask (va_op_xmm_xmm 9) (va_op_reg_opr64_reg64 rR12)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 505 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Pshufb (va_op_xmm_xmm 11) (va_op_xmm_xmm 9)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 506 column 18 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_blocks128 alg in128_b out128_b key round_keys keys_b hkeys_b h_LE) (fun\n (va_s:va_state) _ -> let (y_cipher128:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s) in let (va_arg134:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b in let\n (va_arg133:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = auth_in in let\n (va_arg132:Vale.Def.Types_s.quad32) = y_0 in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 508 column 36 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GHash.lemma_ghash_incremental0_append h_LE va_arg132 y_cipher128x6\n y_cipher128 va_arg133 va_arg134) (let auth_in = FStar.Seq.Base.append #quad32 auth_in\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b) in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 512 column 22 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Stack_lemma (va_op_reg64_reg64 rRsp) (offset + 24) Public) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 512 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Add64 (va_op_dst_opr64_reg64 rR14) (va_opr_code_Stack (va_op_reg64_reg64 rRsp)\n (offset + 24) Public)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 513 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_IMul64 (va_op_dst_opr64_reg64 rR14) (va_const_opr64 16)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 514 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rR13) (va_op_reg_opr64_reg64 rRsp) (offset + 64))\n (fun (va_s:va_state) _ -> let (y_inout:Vale.Def.Types_s.quad32) = y_cipher128 in let\n (plain_byte_seq:(seq quad32)) = empty_seq_quad32 in let (cipher_byte_seq:(seq quad32)) =\n empty_seq_quad32 in let (va_arg131:Vale.Def.Types_s.quad32) = va_get_xmm 11 va_s in let\n (va_arg130:(FStar.Seq.Base.seq Vale.Def.Types_s.nat32)) = key in let\n (va_arg129:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = cipher_byte_seq in let\n (va_arg128:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = plain_byte_seq in let\n (va_arg127:Vale.AES.AES_common_s.algorithm) = alg in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 519 column 29 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.gctr_partial_opaque_init va_arg127 va_arg128 va_arg129 va_arg130\n va_arg131) (let (total_bytes:(va_int_at_least 0)) = FStar.Seq.Base.length #quad32 auth_quad_seq\n `op_Multiply` 16 + plain_num_bytes in va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 523 column 8 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_qIf va_mods (Cmp_gt (va_op_cmp_reg64 rR13) (va_op_cmp_reg64 rR14)) (qblock va_mods (fun\n (va_s:va_state) -> va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 525 column 21 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp) (offset + 56))\n (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 526 column 14 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR10) (va_op_opr64_reg64 rR13)) (fun (va_s:va_state) _\n -> va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 527 column 26 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.Poly1305.Math.lemma_poly_bits64 ()) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 528 column 14 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_And64 (va_op_dst_opr64_reg64 rR10) (va_const_opr64 15)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 532 column 24 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b total_bytes y_0 auth_in\n h_LE) (fun (va_s:va_state) _ -> let y_inout = Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm\n 8 va_s) in let (raw_auth_quads:(FStar.Seq.Base.seq quad32)) = FStar.Seq.Base.append #quad32\n auth_in (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_old_s) inout_b) in va_qAssertSquash\n va_range1\n \"***** EXPRESSION PRECONDITIONS NOT MET WITHIN line 536 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n ((fun a_1906 (s_1907:(FStar.Seq.Base.seq a_1906)) (i_1908:Prims.nat) (j_1909:Prims.nat) -> let\n (j_1869:Prims.nat) = j_1909 in Prims.b2t (Prims.op_AmpAmp (Prims.op_LessThanOrEqual i_1908\n j_1869) (Prims.op_LessThanOrEqual j_1869 (FStar.Seq.Base.length #a_1906 s_1907))))\n Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_auth_quads) 0 total_bytes)\n (fun _ -> let (auth_input_bytes:(FStar.Seq.Base.seq Vale.Def.Types_s.nat8)) =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n raw_auth_quads) 0 total_bytes in let (padded_auth_bytes:(FStar.Seq.Base.seq\n Vale.Def.Types_s.nat8)) = Vale.AES.GCTR_s.pad_to_128_bits auth_input_bytes in let auth_in =\n Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_auth_bytes in let plain_byte_seq =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_old_s) inout_b in let cipher_byte_seq =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s) inout_b in va_QEmpty ((auth_in,\n cipher_byte_seq, plain_byte_seq, y_inout)))))))))) (qblock va_mods (fun (va_s:va_state) ->\n va_QEmpty ((auth_in, cipher_byte_seq, plain_byte_seq, y_inout))))) (fun (va_s:va_state) va_g ->\n let ((auth_in:(seq quad32)), (cipher_byte_seq:(seq quad32)), (plain_byte_seq:(seq quad32)),\n (y_inout:Vale.Def.Types_s.quad32)) = va_g in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 547 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR11) (va_op_opr64_reg64 rR15)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 548 column 25 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_make_length_quad ()) (fun (va_s:va_state) _ -> let\n (length_quad32:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 0\n va_s) in va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 551 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Ghash_register hkeys_b h_LE y_inout) (fun (va_s:va_state) _ -> let\n (y_final:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s)\n in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 554 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64\n rRbp) 0 Secret scratch_b 0) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 557 column 18 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gctr_register alg key round_keys keys_b) (fun (va_s:va_state) _ -> let\n (va_arg126:Vale.Def.Types_s.quad32) = va_get_xmm 8 va_s in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 560 column 40 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.Arch.Types.le_seq_quad32_to_bytes_of_singleton va_arg126)\n (va_qAssertSquash va_range1\n \"***** EXPRESSION PRECONDITIONS NOT MET WITHIN line 561 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n ((fun (icb_BE_677:Vale.Def.Types_s.quad32) (plain_LE_678:Vale.Def.Types_s.quad32)\n (alg_679:Vale.AES.AES_common_s.algorithm) (key_680:(FStar.Seq.Base.seq Vale.Def.Types_s.nat32))\n (i_681:Prims.int) -> Vale.AES.AES_s.is_aes_key_LE alg_679 key_680) j0 y_final alg key 0) (fun _\n -> va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 561 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_xmm 8 va_s == Vale.AES.GCTR_s.gctr_encrypt_block j0 y_final alg key 0) (let\n (plain128:(FStar.Seq.Base.seq Vale.X64.Decls.quad32)) = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b) in let\n (cipher128:(FStar.Seq.Base.seq Vale.X64.Decls.quad32)) = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) in128_b) in va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 566 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (FStar.Seq.Base.length #quad32 plain_byte_seq == 0 ==> FStar.Seq.Base.equal\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 plain128 plain_byte_seq)\n plain128) (va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 567 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (FStar.Seq.Base.length #quad32 cipher_byte_seq == 0 ==> FStar.Seq.Base.equal\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 cipher128 cipher_byte_seq)\n cipher128) (let (va_arg125:Vale.Def.Types_s.quad32) = Vale.AES.GCTR.inc32lite ctr_BE_2 len128x6\n in let (va_arg124:Vale.Def.Types_s.quad32) = ctr_BE_2 in let (va_arg123:(FStar.Seq.Base.seq\n Vale.Def.Types_s.nat32)) = key in let (va_arg122:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32))\n = Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) out128_b in let\n (va_arg121:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_old_s) in128_b in let (va_arg120:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s) out128x6_b in let\n (va_arg119:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_old_s) in128x6_b in let (va_arg118:Prims.nat) = len128 in let\n (va_arg117:Prims.nat) = len128x6 in let (va_arg116:Vale.AES.AES_common_s.algorithm) = alg in\n va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 569 column 30 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.lemma_gctr_partial_append va_arg116 va_arg117 va_arg118\n va_arg119 va_arg120 va_arg121 va_arg122 va_arg123 va_arg124 va_arg125) (let\n (va_arg115:Vale.Def.Types_s.quad32) = Vale.AES.GCTR.inc32lite (Vale.AES.GCTR.inc32lite ctr_BE_2\n len128x6) len128 in let (va_arg114:Vale.Def.Types_s.quad32) = ctr_BE_2 in let\n (va_arg113:(FStar.Seq.Base.seq Vale.Def.Types_s.nat32)) = key in let\n (va_arg112:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = cipher_byte_seq in let\n (va_arg111:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = plain_byte_seq in let\n (va_arg110:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s) out128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) out128_b) in let\n (va_arg109:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b) in let (va_arg108:Prims.nat) =\n FStar.Seq.Base.length #quad32 plain_byte_seq in let (va_arg107:Prims.nat) = len128x6 + len128\n in let (va_arg106:Vale.AES.AES_common_s.algorithm) = alg in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 575 column 30 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.lemma_gctr_partial_append va_arg106 va_arg107 va_arg108\n va_arg109 va_arg110 va_arg111 va_arg112 va_arg113 va_arg114 va_arg115) (let\n (va_arg105:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = auth_in in let\n (va_arg104:Vale.Def.Types_s.quad32) = y_0 in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 583 column 23 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GHash.lemma_hash_append2 h_LE va_arg104 y_inout y_final va_arg105\n length_quad32) (let auth_in = FStar.Seq.Base.append #quad32 auth_in (FStar.Seq.Base.create\n #Vale.Def.Types_s.quad32 1 length_quad32) in let (va_arg103:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = auth_in in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 585 column 31 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GHash.ghash_incremental_to_ghash h_LE va_arg103) (va_QEmpty\n (()))))))))))))))))))))))))))))))))))))))))))))))))\n\n\nval va_lemma_Gcm_blocks : va_b0:va_code -> va_s0:va_state -> alg:algorithm -> offset:int ->\n auth_b:buffer128 -> abytes_b:buffer128 -> in128x6_b:buffer128 -> out128x6_b:buffer128 ->\n in128_b:buffer128 -> out128_b:buffer128 -> inout_b:buffer128 -> iv_b:buffer128 ->\n scratch_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_blocks alg offset) va_s0 /\\ va_get_ok va_s0 /\\\n (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in let (h_LE:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.X64.Decls.buffer128_read hkeys_b 2\n (va_get_mem_heaplet 0 va_s0)) in sse_enabled /\\ movbe_enabled /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0)\n Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp\n va_s0 + offset + 8) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 24) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 40) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 56) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet\n 1 va_s0) (va_get_reg64 rRdi va_s0) auth_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0)\n Secret /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 7 va_s0) abytes_ptr abytes_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 2\n va_s0) (va_get_reg64 rR8 va_s0) iv_b 1 (va_get_mem_layout va_s0) Public /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 6 va_s0) in128x6_ptr in128x6_b len128x6\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 6\n va_s0) out128x6_ptr out128x6_b len128x6 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0) in128_ptr in128_b len128\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1\n va_s0) out128_ptr out128_b len128 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) inout_ptr inout_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 3\n va_s0) (va_get_reg64 rRbp va_s0) scratch_b 9 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0) hkeys_b 8\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.buffer_disjoints128 iv_b ([keys_b;\n scratch_b; in128x6_b; out128x6_b; hkeys_b; in128_b; out128_b; inout_b]) /\\\n Vale.X64.Decls.buffer_disjoints128 scratch_b ([keys_b; in128x6_b; out128x6_b; in128_b;\n out128_b; inout_b; hkeys_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128x6_b ([keys_b;\n hkeys_b; in128_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128_b ([keys_b; hkeys_b;\n out128x6_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 inout_b ([keys_b; hkeys_b;\n out128x6_b; out128_b]) /\\ (Vale.X64.Decls.buffers_disjoint128 in128x6_b out128x6_b \\/ in128x6_b\n == out128x6_b) /\\ (Vale.X64.Decls.buffers_disjoint128 in128_b out128_b \\/ in128_b == out128_b)\n /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ in128x6_ptr\n + 16 `op_Multiply` len128x6 < pow2_64 /\\ out128x6_ptr + 16 `op_Multiply` len128x6 < pow2_64 /\\\n in128_ptr + 16 `op_Multiply` len128 < pow2_64 /\\ out128_ptr + 16 `op_Multiply` len128 < pow2_64\n /\\ inout_ptr + 16 < pow2_64 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 auth_b ==\n va_get_reg64 rRdx va_s0 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 abytes_b == 1\n /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128x6_b ==\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out128x6_b /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 out128_b /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128\n in128x6_b == len128x6 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b ==\n len128 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 inout_b == 1 /\\\n plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32 /\\ va_get_reg64 rR9 va_s0 + 32 <\n pow2_64 /\\ Vale.X64.Memory.buffer_addr #Vale.X64.Memory.vuint128 keys_b (va_get_mem_heaplet 0\n va_s0) + 128 < pow2_64 /\\ len128x6 `op_Modulus` 6 == 0 /\\ (len128x6 > 0 ==> len128x6 >= 6) /\\\n 12 + len128x6 + 6 < pow2_32 /\\ (va_mul_nat len128x6 (128 `op_Division` 8) + va_mul_nat len128\n (128 `op_Division` 8) <= plain_num_bytes /\\ plain_num_bytes < va_mul_nat len128x6 (128\n `op_Division` 8) + va_mul_nat len128 (128 `op_Division` 8) + 128 `op_Division` 8) /\\\n (va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) <= va_get_reg64 rRsi va_s0 /\\\n va_get_reg64 rRsi va_s0 < va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) + 128\n `op_Division` 8) /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rRcx va_s0)\n (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 (Vale.AES.AES_s.aes_encrypt_LE alg key\n (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0))))))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in let (h_LE:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.X64.Decls.buffer128_read hkeys_b 2\n (va_get_mem_heaplet 0 va_s0)) in Vale.X64.Decls.modifies_buffer128 out128_b (va_get_mem_heaplet\n 1 va_s0) (va_get_mem_heaplet 1 va_sM) /\\ Vale.X64.Decls.modifies_buffer128 iv_b\n (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2 va_sM) /\\ Vale.X64.Decls.modifies_buffer128\n scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3 va_sM) /\\\n Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM) /\\ Vale.X64.Decls.modifies_buffer128 out128x6_b (va_get_mem_heaplet 6 va_s0)\n (va_get_mem_heaplet 6 va_sM) /\\ plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32\n /\\ (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in let\n (ctr_BE_1:quad32) = iv_BE in let (ctr_BE_2:quad32) = Vale.AES.GCTR_s.inc32 iv_BE 1 in let\n (plain_in:(seq quad32)) = (if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) then FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b) else FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in128_b)) in let (cipher_out:(seq quad32)) = (if (plain_num_bytes\n > (len128x6 + len128) `op_Multiply` 128 `op_Division` 8) then FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM)\n out128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM) inout_b) else\n FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM)\n out128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out128_b)) in let\n (cipher_bound:nat) = (if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128 `op_Division`\n 8) then (len128x6 + len128 + 1) else (len128x6 + len128)) in Vale.AES.GCTR.gctr_partial alg\n cipher_bound plain_in cipher_out key ctr_BE_2 /\\ (let (length_quad:quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.Def.Types_s.insert_nat64\n (Vale.Def.Types_s.insert_nat64 (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0) (8\n `op_Multiply` va_get_reg64 rRsi va_s0) 1) (8 `op_Multiply` plain_num_bytes) 0) in let\n (raw_auth_quads:(seq quad32)) = (if (va_get_reg64 rRsi va_s0 > va_get_reg64 rRdx va_s0\n `op_Multiply` 128 `op_Division` 8) then FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 7 va_s0) abytes_b) else Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0)\n auth_b) in let (auth_input_bytes:(seq nat8)) = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_auth_quads) 0 (va_get_reg64 rRsi va_s0) in let\n (padded_auth_bytes:(seq nat8)) = Vale.AES.GCTR_s.pad_to_128_bits auth_input_bytes in let\n (auth_quad_seq:(seq quad32)) = Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_auth_bytes in let\n (raw_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32 (FStar.Seq.Base.append #quad32\n auth_quad_seq (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b))\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b) in let (total_bytes:nat) =\n FStar.Seq.Base.length #quad32 auth_quad_seq `op_Multiply` 16 + plain_num_bytes in let\n (raw_quad_seq:(seq quad32)) = (if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) then (let (ab:(seq nat8)) = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes (FStar.Seq.Base.append #quad32 raw_quad_seq\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b))) 0 total_bytes in let (pb:(seq\n nat8)) = Vale.AES.GCTR_s.pad_to_128_bits ab in Vale.Def.Types_s.le_bytes_to_seq_quad32 pb) else\n raw_quad_seq) in let (auth_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32 raw_quad_seq\n (FStar.Seq.Base.create #quad32 1 length_quad) in va_get_xmm 8 va_sM ==\n Vale.AES.GCTR_s.gctr_encrypt_block ctr_BE_1 (Vale.AES.GHash_s.ghash_LE h_LE auth_quad_seq) alg\n key 0))) /\\ va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 6 va_sM\n (va_update_mem_heaplet 5 va_sM (va_update_mem_heaplet 3 va_sM (va_update_mem_heaplet 2 va_sM\n (va_update_mem_heaplet 1 va_sM (va_update_xmm 15 va_sM (va_update_xmm 14 va_sM (va_update_xmm\n 13 va_sM (va_update_xmm 12 va_sM (va_update_xmm 11 va_sM (va_update_xmm 10 va_sM (va_update_xmm\n 9 va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5\n va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1\n va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR15 va_sM (va_update_reg64 rR14 va_sM\n (va_update_reg64 rR13 va_sM (va_update_reg64 rR12 va_sM (va_update_reg64 rR11 va_sM\n (va_update_reg64 rR10 va_sM (va_update_reg64 rR9 va_sM (va_update_reg64 rR8 va_sM\n (va_update_reg64 rRbp va_sM (va_update_reg64 rRsi va_sM (va_update_reg64 rRdi va_sM\n (va_update_reg64 rRdx va_sM (va_update_reg64 rRcx va_sM (va_update_reg64 rRbx va_sM\n (va_update_reg64 rRax va_sM (va_update_ok va_sM (va_update_mem va_sM\n va_s0)))))))))))))))))))))))))))))))))))))))))\n[@\"opaque_to_smt\"]\nlet va_lemma_Gcm_blocks va_b0 va_s0 alg offset auth_b abytes_b in128x6_b out128x6_b in128_b\n out128_b inout_b iv_b scratch_b key round_keys keys_b hkeys_b =\n let (va_mods:va_mods_t) = [va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5;\n va_Mod_mem_heaplet 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14;\n va_Mod_xmm 13; va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8;\n va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm\n 1; va_Mod_xmm 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12;\n va_Mod_reg64 rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp;\n va_Mod_reg64 rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx;\n va_Mod_reg64 rRax; va_Mod_ok; va_Mod_mem] in\n let va_qc = va_qcode_Gcm_blocks va_mods alg offset auth_b abytes_b in128x6_b out128x6_b in128_b\n out128_b inout_b iv_b scratch_b key round_keys keys_b hkeys_b in\n let (va_sM, va_fM, va_g) = va_wp_sound_code_norm (va_code_Gcm_blocks alg offset) va_qc va_s0 (fun\n va_s0 va_sM va_g -> let () = va_g in label va_range1\n \"***** POSTCONDITION NOT MET AT line 283 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\ (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64\n (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0) in let\n (in128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 8) (va_get_stack va_s0) in let (out128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack va_s0) in\n let (len128x6:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 24) (va_get_stack va_s0) in let (in128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack va_s0) in\n let (out128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 40) (va_get_stack va_s0) in let (len128:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack va_s0) in\n let (inout_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 56) (va_get_stack va_s0) in let (plain_num_bytes:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack va_s0) in\n let (h_LE:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.X64.Decls.buffer128_read hkeys_b 2 (va_get_mem_heaplet 0 va_s0)) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 396 column 56 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out128_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 397 column 52 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 iv_b (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 398 column 57 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 399 column 55 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 400 column 58 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out128x6_b (va_get_mem_heaplet 6 va_s0) (va_get_mem_heaplet\n 6 va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 403 column 39 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (plain_num_bytes < pow2_32) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 404 column 38 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRsi va_s0 < pow2_32) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 406 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 408 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (ctr_BE_1:quad32) = iv_BE in label va_range1\n \"***** POSTCONDITION NOT MET AT line 409 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (ctr_BE_2:quad32) = Vale.AES.GCTR_s.inc32 iv_BE 1 in label va_range1\n \"***** POSTCONDITION NOT MET AT line 412 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (plain_in:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b)) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in128_b)) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 421 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (cipher_out:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM) out128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b)) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_sM) out128_b)) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 430 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (cipher_bound:nat) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> len128x6 + len128 + 1) (fun _ -> len128x6 + len128) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 434 column 77 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.AES.GCTR.gctr_partial alg cipher_bound plain_in cipher_out key ctr_BE_2) /\\ label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 438 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (length_quad:quad32) = Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.Def.Types_s.insert_nat64 (Vale.Def.Types_s.insert_nat64 (Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 0 0 0 0) (8 `op_Multiply` va_get_reg64 rRsi va_s0) 1) (8 `op_Multiply`\n plain_num_bytes) 0) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 440 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (raw_auth_quads:(seq quad32)) = va_if (va_get_reg64 rRsi va_s0 > va_get_reg64 rRdx va_s0\n `op_Multiply` 128 `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 7 va_s0) abytes_b)) (fun _ -> Vale.X64.Decls.s128 (va_get_mem_heaplet 1\n va_s0) auth_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 444 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (auth_input_bytes:(seq nat8)) = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_auth_quads) 0 (va_get_reg64 rRsi va_s0) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 445 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (padded_auth_bytes:(seq nat8)) = Vale.AES.GCTR_s.pad_to_128_bits auth_input_bytes in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 446 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (auth_quad_seq:(seq quad32)) = Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_auth_bytes\n in label va_range1\n \"***** POSTCONDITION NOT MET AT line 448 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (raw_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32 (FStar.Seq.Base.append #quad32\n auth_quad_seq (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b))\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 452 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (total_bytes:nat) = FStar.Seq.Base.length #quad32 auth_quad_seq `op_Multiply` 16 +\n plain_num_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 453 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (raw_quad_seq:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply`\n 128 `op_Division` 8) (fun _ -> let (ab:(seq nat8)) = FStar.Seq.Base.slice\n #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes (FStar.Seq.Base.append #quad32\n raw_quad_seq (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b))) 0 total_bytes in let\n (pb:(seq nat8)) = Vale.AES.GCTR_s.pad_to_128_bits ab in Vale.Def.Types_s.le_bytes_to_seq_quad32\n pb) (fun _ -> raw_quad_seq) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 460 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (auth_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32 raw_quad_seq\n (FStar.Seq.Base.create #quad32 1 length_quad) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 461 column 106 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_xmm 8 va_sM == Vale.AES.GCTR_s.gctr_encrypt_block ctr_BE_1 (Vale.AES.GHash_s.ghash_LE\n h_LE auth_quad_seq) alg key 0)))))))))))))))))) in\n assert_norm (va_qc.mods == va_mods);\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet\n 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13;\n va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm\n 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64\n rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64\n rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64\n rRax; va_Mod_ok; va_Mod_mem]) va_sM va_s0;\n (va_sM, va_fM)\n\n[@ va_qattr]\nlet va_wp_Gcm_blocks (alg:algorithm) (offset:int) (auth_b:buffer128) (abytes_b:buffer128)\n (in128x6_b:buffer128) (out128x6_b:buffer128) (in128_b:buffer128) (out128_b:buffer128)\n (inout_b:buffer128) (iv_b:buffer128) (scratch_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (va_s0:va_state) (va_k:(va_state -> unit ->\n Type0)) : Type0 =\n (va_get_ok va_s0 /\\ (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64\n (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0) in let\n (in128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 8) (va_get_stack va_s0) in let (out128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack va_s0) in\n let (len128x6:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 24) (va_get_stack va_s0) in let (in128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack va_s0) in\n let (out128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 40) (va_get_stack va_s0) in let (len128:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack va_s0) in\n let (inout_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 56) (va_get_stack va_s0) in let (plain_num_bytes:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack va_s0) in\n let (h_LE:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.X64.Decls.buffer128_read hkeys_b 2 (va_get_mem_heaplet 0 va_s0)) in sse_enabled /\\\n movbe_enabled /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 0)\n (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64\n (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 24) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 40) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 56) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet\n 1 va_s0) (va_get_reg64 rRdi va_s0) auth_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0)\n Secret /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 7 va_s0) abytes_ptr abytes_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 2\n va_s0) (va_get_reg64 rR8 va_s0) iv_b 1 (va_get_mem_layout va_s0) Public /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 6 va_s0) in128x6_ptr in128x6_b len128x6\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 6\n va_s0) out128x6_ptr out128x6_b len128x6 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0) in128_ptr in128_b len128\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1\n va_s0) out128_ptr out128_b len128 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) inout_ptr inout_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 3\n va_s0) (va_get_reg64 rRbp va_s0) scratch_b 9 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0) hkeys_b 8\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.buffer_disjoints128 iv_b ([keys_b;\n scratch_b; in128x6_b; out128x6_b; hkeys_b; in128_b; out128_b; inout_b]) /\\\n Vale.X64.Decls.buffer_disjoints128 scratch_b ([keys_b; in128x6_b; out128x6_b; in128_b;\n out128_b; inout_b; hkeys_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128x6_b ([keys_b;\n hkeys_b; in128_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128_b ([keys_b; hkeys_b;\n out128x6_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 inout_b ([keys_b; hkeys_b;\n out128x6_b; out128_b]) /\\ (Vale.X64.Decls.buffers_disjoint128 in128x6_b out128x6_b \\/ in128x6_b\n == out128x6_b) /\\ (Vale.X64.Decls.buffers_disjoint128 in128_b out128_b \\/ in128_b == out128_b)\n /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ in128x6_ptr\n + 16 `op_Multiply` len128x6 < pow2_64 /\\ out128x6_ptr + 16 `op_Multiply` len128x6 < pow2_64 /\\\n in128_ptr + 16 `op_Multiply` len128 < pow2_64 /\\ out128_ptr + 16 `op_Multiply` len128 < pow2_64\n /\\ inout_ptr + 16 < pow2_64 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 auth_b ==\n va_get_reg64 rRdx va_s0 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 abytes_b == 1\n /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128x6_b ==\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out128x6_b /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 out128_b /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128\n in128x6_b == len128x6 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b ==\n len128 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 inout_b == 1 /\\\n plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32 /\\ va_get_reg64 rR9 va_s0 + 32 <\n pow2_64 /\\ Vale.X64.Memory.buffer_addr #Vale.X64.Memory.vuint128 keys_b (va_get_mem_heaplet 0\n va_s0) + 128 < pow2_64 /\\ len128x6 `op_Modulus` 6 == 0 /\\ (len128x6 > 0 ==> len128x6 >= 6) /\\\n 12 + len128x6 + 6 < pow2_32 /\\ (va_mul_nat len128x6 (128 `op_Division` 8) + va_mul_nat len128\n (128 `op_Division` 8) <= plain_num_bytes /\\ plain_num_bytes < va_mul_nat len128x6 (128\n `op_Division` 8) + va_mul_nat len128 (128 `op_Division` 8) + 128 `op_Division` 8) /\\\n (va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) <= va_get_reg64 rRsi va_s0 /\\\n va_get_reg64 rRsi va_s0 < va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) + 128\n `op_Division` 8) /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rRcx va_s0)\n (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 (Vale.AES.AES_s.aes_encrypt_LE alg key\n (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0)))) /\\ (forall (va_x_mem:vale_heap)\n (va_x_rax:nat64) (va_x_rbx:nat64) (va_x_rcx:nat64) (va_x_rdx:nat64) (va_x_rdi:nat64)\n (va_x_rsi:nat64) (va_x_rbp:nat64) (va_x_r8:nat64) (va_x_r9:nat64) (va_x_r10:nat64)\n (va_x_r11:nat64) (va_x_r12:nat64) (va_x_r13:nat64) (va_x_r14:nat64) (va_x_r15:nat64)\n (va_x_xmm0:quad32) (va_x_xmm1:quad32) (va_x_xmm2:quad32) (va_x_xmm3:quad32) (va_x_xmm4:quad32)\n (va_x_xmm5:quad32) (va_x_xmm6:quad32) (va_x_xmm7:quad32) (va_x_xmm8:quad32) (va_x_xmm9:quad32)\n (va_x_xmm10:quad32) (va_x_xmm11:quad32) (va_x_xmm12:quad32) (va_x_xmm13:quad32)\n (va_x_xmm14:quad32) (va_x_xmm15:quad32) (va_x_heap1:vale_heap) (va_x_heap2:vale_heap)\n (va_x_heap3:vale_heap) (va_x_heap5:vale_heap) (va_x_heap6:vale_heap)\n (va_x_efl:Vale.X64.Flags.t) . let va_sM = va_upd_flags va_x_efl (va_upd_mem_heaplet 6\n va_x_heap6 (va_upd_mem_heaplet 5 va_x_heap5 (va_upd_mem_heaplet 3 va_x_heap3\n (va_upd_mem_heaplet 2 va_x_heap2 (va_upd_mem_heaplet 1 va_x_heap1 (va_upd_xmm 15 va_x_xmm15\n (va_upd_xmm 14 va_x_xmm14 (va_upd_xmm 13 va_x_xmm13 (va_upd_xmm 12 va_x_xmm12 (va_upd_xmm 11\n va_x_xmm11 (va_upd_xmm 10 va_x_xmm10 (va_upd_xmm 9 va_x_xmm9 (va_upd_xmm 8 va_x_xmm8\n (va_upd_xmm 7 va_x_xmm7 (va_upd_xmm 6 va_x_xmm6 (va_upd_xmm 5 va_x_xmm5 (va_upd_xmm 4 va_x_xmm4\n (va_upd_xmm 3 va_x_xmm3 (va_upd_xmm 2 va_x_xmm2 (va_upd_xmm 1 va_x_xmm1 (va_upd_xmm 0 va_x_xmm0\n (va_upd_reg64 rR15 va_x_r15 (va_upd_reg64 rR14 va_x_r14 (va_upd_reg64 rR13 va_x_r13\n (va_upd_reg64 rR12 va_x_r12 (va_upd_reg64 rR11 va_x_r11 (va_upd_reg64 rR10 va_x_r10\n (va_upd_reg64 rR9 va_x_r9 (va_upd_reg64 rR8 va_x_r8 (va_upd_reg64 rRbp va_x_rbp (va_upd_reg64\n rRsi va_x_rsi (va_upd_reg64 rRdi va_x_rdi (va_upd_reg64 rRdx va_x_rdx (va_upd_reg64 rRcx\n va_x_rcx (va_upd_reg64 rRbx va_x_rbx (va_upd_reg64 rRax va_x_rax (va_upd_mem va_x_mem\n va_s0))))))))))))))))))))))))))))))))))))) in va_get_ok va_sM /\\ (let\n (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in let (h_LE:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.X64.Decls.buffer128_read hkeys_b 2\n (va_get_mem_heaplet 0 va_s0)) in Vale.X64.Decls.modifies_buffer128 out128_b (va_get_mem_heaplet\n 1 va_s0) (va_get_mem_heaplet 1 va_sM) /\\ Vale.X64.Decls.modifies_buffer128 iv_b\n (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2 va_sM) /\\ Vale.X64.Decls.modifies_buffer128\n scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3 va_sM) /\\\n Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM) /\\ Vale.X64.Decls.modifies_buffer128 out128x6_b (va_get_mem_heaplet 6 va_s0)\n (va_get_mem_heaplet 6 va_sM) /\\ plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32\n /\\ (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in let\n (ctr_BE_1:quad32) = iv_BE in let (ctr_BE_2:quad32) = Vale.AES.GCTR_s.inc32 iv_BE 1 in let\n (plain_in:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b)) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in128_b)) in let (cipher_out:(seq quad32)) = va_if\n (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128 `op_Division` 8) (fun _ ->\n FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_sM) out128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM)\n inout_b)) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM)\n out128_b)) in let (cipher_bound:nat) = va_if (plain_num_bytes > (len128x6 + len128)\n `op_Multiply` 128 `op_Division` 8) (fun _ -> len128x6 + len128 + 1) (fun _ -> len128x6 +\n len128) in Vale.AES.GCTR.gctr_partial alg cipher_bound plain_in cipher_out key ctr_BE_2 /\\ (let\n (length_quad:quad32) = Vale.Def.Types_s.reverse_bytes_quad32 (Vale.Def.Types_s.insert_nat64\n (Vale.Def.Types_s.insert_nat64 (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0) (8\n `op_Multiply` va_get_reg64 rRsi va_s0) 1) (8 `op_Multiply` plain_num_bytes) 0) in let\n (raw_auth_quads:(seq quad32)) = va_if (va_get_reg64 rRsi va_s0 > va_get_reg64 rRdx va_s0\n `op_Multiply` 128 `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 7 va_s0) abytes_b)) (fun _ -> Vale.X64.Decls.s128 (va_get_mem_heaplet 1\n va_s0) auth_b) in let (auth_input_bytes:(seq nat8)) = FStar.Seq.Base.slice\n #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_auth_quads) 0 (va_get_reg64\n rRsi va_s0) in let (padded_auth_bytes:(seq nat8)) = Vale.AES.GCTR_s.pad_to_128_bits\n auth_input_bytes in let (auth_quad_seq:(seq quad32)) = Vale.Def.Types_s.le_bytes_to_seq_quad32\n padded_auth_bytes in let (raw_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32\n (FStar.Seq.Base.append #quad32 auth_quad_seq (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0)\n in128x6_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b) in let (total_bytes:nat)\n = FStar.Seq.Base.length #quad32 auth_quad_seq `op_Multiply` 16 + plain_num_bytes in let\n (raw_quad_seq:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> let (ab:(seq nat8)) = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes (FStar.Seq.Base.append #quad32 raw_quad_seq\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b))) 0 total_bytes in let (pb:(seq\n nat8)) = Vale.AES.GCTR_s.pad_to_128_bits ab in Vale.Def.Types_s.le_bytes_to_seq_quad32 pb) (fun\n _ -> raw_quad_seq) in let (auth_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32\n raw_quad_seq (FStar.Seq.Base.create #quad32 1 length_quad) in va_get_xmm 8 va_sM ==\n Vale.AES.GCTR_s.gctr_encrypt_block ctr_BE_1 (Vale.AES.GHash_s.ghash_LE h_LE auth_quad_seq) alg\n key 0))) ==> va_k va_sM (())))\n\n\nval va_wpProof_Gcm_blocks : alg:algorithm -> offset:int -> auth_b:buffer128 -> abytes_b:buffer128\n -> in128x6_b:buffer128 -> out128x6_b:buffer128 -> in128_b:buffer128 -> out128_b:buffer128 ->\n inout_b:buffer128 -> iv_b:buffer128 -> scratch_b:buffer128 -> key:(seq nat32) -> round_keys:(seq\n quad32) -> keys_b:buffer128 -> hkeys_b:buffer128 -> va_s0:va_state -> va_k:(va_state -> unit ->\n Type0)\n -> Ghost (va_state & va_fuel & unit)\n (requires (va_t_require va_s0 /\\ va_wp_Gcm_blocks alg offset auth_b abytes_b in128x6_b out128x6_b\n in128_b out128_b inout_b iv_b scratch_b key round_keys keys_b hkeys_b va_s0 va_k))\n (ensures (fun (va_sM, va_f0, va_g) -> va_t_ensure (va_code_Gcm_blocks alg offset) ([va_Mod_flags;\n va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet 3; va_Mod_mem_heaplet 2;\n va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13; va_Mod_xmm 12; va_Mod_xmm\n 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5;\n va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR15;\n va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64 rR11; va_Mod_reg64 rR10;\n va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64 rRsi; va_Mod_reg64 rRdi;\n va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64 rRax; va_Mod_mem]) va_s0\n va_k ((va_sM, va_f0, va_g))))\n[@\"opaque_to_smt\"]\nlet va_wpProof_Gcm_blocks alg offset auth_b abytes_b in128x6_b out128x6_b in128_b out128_b inout_b\n iv_b scratch_b key round_keys keys_b hkeys_b va_s0 va_k =\n let (va_sM, va_f0) = va_lemma_Gcm_blocks (va_code_Gcm_blocks alg offset) va_s0 alg offset auth_b\n abytes_b in128x6_b out128x6_b in128_b out128_b inout_b iv_b scratch_b key round_keys keys_b\n hkeys_b in\n va_lemma_upd_update va_sM;\n assert (va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 6 va_sM\n (va_update_mem_heaplet 5 va_sM (va_update_mem_heaplet 3 va_sM (va_update_mem_heaplet 2 va_sM\n (va_update_mem_heaplet 1 va_sM (va_update_xmm 15 va_sM (va_update_xmm 14 va_sM (va_update_xmm\n 13 va_sM (va_update_xmm 12 va_sM (va_update_xmm 11 va_sM (va_update_xmm 10 va_sM (va_update_xmm\n 9 va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5\n va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1\n va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR15 va_sM (va_update_reg64 rR14 va_sM\n (va_update_reg64 rR13 va_sM (va_update_reg64 rR12 va_sM (va_update_reg64 rR11 va_sM\n (va_update_reg64 rR10 va_sM (va_update_reg64 rR9 va_sM (va_update_reg64 rR8 va_sM\n (va_update_reg64 rRbp va_sM (va_update_reg64 rRsi va_sM (va_update_reg64 rRdi va_sM\n (va_update_reg64 rRdx va_sM (va_update_reg64 rRcx va_sM (va_update_reg64 rRbx va_sM\n (va_update_reg64 rRax va_sM (va_update_ok va_sM (va_update_mem va_sM\n va_s0))))))))))))))))))))))))))))))))))))))));\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet\n 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13;\n va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm\n 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64\n rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64\n rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64\n rRax; va_Mod_mem]) va_sM va_s0;\n let va_g = () in\n (va_sM, va_f0, va_g)\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_quick_Gcm_blocks (alg:algorithm) (offset:int) (auth_b:buffer128) (abytes_b:buffer128)\n (in128x6_b:buffer128) (out128x6_b:buffer128) (in128_b:buffer128) (out128_b:buffer128)\n (inout_b:buffer128) (iv_b:buffer128) (scratch_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) : (va_quickCode unit (va_code_Gcm_blocks alg\n offset)) =\n (va_QProc (va_code_Gcm_blocks alg offset) ([va_Mod_flags; va_Mod_mem_heaplet 6;\n va_Mod_mem_heaplet 5; va_Mod_mem_heaplet 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1;\n va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13; va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10;\n va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm\n 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64\n rR13; va_Mod_reg64 rR12; va_Mod_reg64 rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64\n rR8; va_Mod_reg64 rRbp; va_Mod_reg64 rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64\n rRcx; va_Mod_reg64 rRbx; va_Mod_reg64 rRax; va_Mod_mem]) (va_wp_Gcm_blocks alg offset auth_b\n abytes_b in128x6_b out128x6_b in128_b out128_b inout_b iv_b scratch_b key round_keys keys_b\n hkeys_b) (va_wpProof_Gcm_blocks alg offset auth_b abytes_b in128x6_b out128x6_b in128_b\n out128_b inout_b iv_b scratch_b key round_keys keys_b hkeys_b))\n#pop-options\n//--\n//-- Gcm_blocks_wrapped\n#push-options \"--z3rlimit 60\"\n\nval va_code_Gcm_blocks_wrapped : alg:algorithm -> offset:int -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_blocks_wrapped alg offset =\n (va_Block (va_CCons (va_code_Gcm_blocks alg offset) (va_CCons (va_Block (va_CNil ())) (va_CCons\n (va_Block (va_CNil ())) (va_CNil ())))))\n\nval va_codegen_success_Gcm_blocks_wrapped : alg:algorithm -> offset:int -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_blocks_wrapped alg offset =\n (va_pbool_and (va_codegen_success_Gcm_blocks alg offset) (va_ttrue ()))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_blocks_wrapped (va_mods:va_mods_t) (alg:algorithm) (offset:int) (auth_b:buffer128)\n (abytes_b:buffer128) (in128x6_b:buffer128) (out128x6_b:buffer128) (in128_b:buffer128)\n (out128_b:buffer128) (inout_b:buffer128) (iv_b:buffer128) (iv:supported_iv_LE)\n (scratch_b:buffer128) (key:(seq nat32)) (round_keys:(seq quad32)) (keys_b:buffer128)\n (hkeys_b:buffer128) (expected_tag:(seq nat8)) : (va_quickCode unit (va_code_Gcm_blocks_wrapped\n alg offset)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in let\n (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 0) (va_get_stack va_s) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 8) (va_get_stack va_s) in let\n (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 16) (va_get_stack va_s) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 24) (va_get_stack va_s) in let\n (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 32) (va_get_stack va_s) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 40) (va_get_stack va_s) in let\n (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset\n + 48) (va_get_stack va_s) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 56) (va_get_stack va_s) in let\n (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s\n + offset + 64) (va_get_stack va_s) in va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 739 column 15 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_blocks alg offset auth_b abytes_b in128x6_b out128x6_b in128_b out128_b inout_b\n iv_b scratch_b key round_keys keys_b hkeys_b) (fun (va_s:va_state) _ -> va_qAssertSquash\n va_range1\n \"***** EXPRESSION PRECONDITIONS NOT MET WITHIN line 741 column 37 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (8 `op_Multiply` va_get_reg64 rRsi va_old_s >= 0 /\\ 8 `op_Multiply` va_get_reg64 rRsi va_old_s\n <= 18446744073709551615 /\\ 8 `op_Multiply` plain_num_bytes >= 0 /\\ 8 `op_Multiply`\n plain_num_bytes <= 18446744073709551615) (fun _ -> let (va_arg55:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.Def.Types_s.insert_nat64\n (Vale.Def.Types_s.insert_nat64 (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0) (8\n `op_Multiply` va_get_reg64 rRsi va_old_s) 1) (8 `op_Multiply` plain_num_bytes) 0) in let\n (va_arg54:Vale.Def.Types_s.quad32) = va_get_xmm 8 va_s in let\n (va_arg53:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.X64.Decls.buffer128_read hkeys_b 2 (va_get_mem_heaplet 0 va_s)) in let\n (va_arg52:Vale.Def.Types_s.quad32) = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2\n va_old_s) in let (va_arg51:Vale.AES.GCM_s.supported_iv_LE) = iv in let (va_arg50:Prims.nat) =\n va_get_reg64 rRsi va_old_s in let (va_arg49:Prims.nat) = plain_num_bytes in let\n (va_arg48:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s) inout_b in let (va_arg47:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) out128_b in let\n (va_arg46:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_s) out128x6_b in let (va_arg45:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_old_s) inout_b in let\n (va_arg44:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_old_s) in128_b in let (va_arg43:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b in\n let (va_arg42:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 7 va_old_s) abytes_b in let (va_arg41:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) auth_b in let\n (va_arg40:(FStar.Seq.Base.seq Vale.Def.Words_s.nat32)) = key in let\n (va_arg39:Vale.AES.AES_common_s.algorithm) = alg in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 741 column 37 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCM.gcm_blocks_dec_helper_simplified va_arg39 va_arg40 va_arg41\n va_arg42 va_arg43 va_arg44 va_arg45 va_arg46 va_arg47 va_arg48 va_arg49 va_arg50 va_arg51\n va_arg52 va_arg53 va_arg54 va_arg55) (let (auth_raw_quads:(FStar.Seq.Base.seq\n Vale.X64.Decls.quad32)) = FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_old_s) auth_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 7 va_old_s)\n abytes_b) in va_qAssertSquash va_range1\n \"***** EXPRESSION PRECONDITIONS NOT MET WITHIN line 751 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n ((fun a_1906 (s_1907:(FStar.Seq.Base.seq a_1906)) (i_1908:Prims.nat) (j_1909:Prims.nat) -> let\n (j_1869:Prims.nat) = j_1909 in Prims.b2t (Prims.op_AmpAmp (Prims.op_LessThanOrEqual i_1908\n j_1869) (Prims.op_LessThanOrEqual j_1869 (FStar.Seq.Base.length #a_1906 s_1907))))\n Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes auth_raw_quads) 0 (va_get_reg64\n rRsi va_old_s)) (fun _ -> let (auth_bytes:(FStar.Seq.Base.seq Vale.Def.Types_s.nat8)) =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n auth_raw_quads) 0 (va_get_reg64 rRsi va_old_s) in let (va_arg38:Vale.Def.Types_s.quad32) =\n Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_old_s) in let\n (va_arg37:Vale.AES.GCM_s.supported_iv_LE) = iv in let (va_arg36:Prims.nat) = plain_num_bytes in\n let (va_arg35:(FStar.Seq.Base.seq Vale.Def.Words_s.nat8)) = expected_tag in let\n (va_arg34:(FStar.Seq.Base.seq Vale.Def.Words_s.nat8)) = auth_bytes in let\n (va_arg33:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s) inout_b in let (va_arg32:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) out128_b in let\n (va_arg31:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_s) out128x6_b in let (va_arg30:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_old_s) inout_b in let\n (va_arg29:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_old_s) in128_b in let (va_arg28:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b in\n let (va_arg27:(FStar.Seq.Base.seq Vale.Def.Words_s.nat32)) = key in let\n (va_arg26:Vale.AES.AES_common_s.algorithm) = alg in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 752 column 37 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCM.gcm_blocks_helper_dec_simplified va_arg26 va_arg27 va_arg28\n va_arg29 va_arg30 va_arg31 va_arg32 va_arg33 va_arg34 va_arg35 va_arg36 va_arg37 va_arg38)\n (va_QEmpty (()))))))))\n\n\nval va_lemma_Gcm_blocks_wrapped : va_b0:va_code -> va_s0:va_state -> alg:algorithm -> offset:int ->\n auth_b:buffer128 -> abytes_b:buffer128 -> in128x6_b:buffer128 -> out128x6_b:buffer128 ->\n in128_b:buffer128 -> out128_b:buffer128 -> inout_b:buffer128 -> iv_b:buffer128 ->\n iv:supported_iv_LE -> scratch_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) ->\n keys_b:buffer128 -> hkeys_b:buffer128 -> expected_tag:(seq nat8)\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_blocks_wrapped alg offset) va_s0 /\\ va_get_ok\n va_s0 /\\ (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64\n rRsp va_s0 + offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in sse_enabled /\\ movbe_enabled /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0)\n Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp\n va_s0 + offset + 8) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 24) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 40) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 56) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet\n 1 va_s0) (va_get_reg64 rRdi va_s0) auth_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0)\n Secret /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 7 va_s0) abytes_ptr abytes_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 2\n va_s0) (va_get_reg64 rR8 va_s0) iv_b 1 (va_get_mem_layout va_s0) Public /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 6 va_s0) in128x6_ptr in128x6_b len128x6\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 6\n va_s0) out128x6_ptr out128x6_b len128x6 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0) in128_ptr in128_b len128\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1\n va_s0) out128_ptr out128_b len128 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) inout_ptr inout_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 3\n va_s0) (va_get_reg64 rRbp va_s0) scratch_b 9 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0) hkeys_b 8\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.buffer_disjoints128 iv_b ([keys_b;\n scratch_b; in128x6_b; out128x6_b; hkeys_b; in128_b; out128_b; inout_b]) /\\\n Vale.X64.Decls.buffer_disjoints128 scratch_b ([keys_b; in128x6_b; out128x6_b; in128_b;\n out128_b; inout_b; hkeys_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128x6_b ([keys_b;\n hkeys_b; in128_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128_b ([keys_b; hkeys_b;\n out128x6_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 inout_b ([keys_b; hkeys_b;\n out128x6_b; out128_b]) /\\ (Vale.X64.Decls.buffers_disjoint128 in128x6_b out128x6_b \\/ in128x6_b\n == out128x6_b) /\\ (Vale.X64.Decls.buffers_disjoint128 in128_b out128_b \\/ in128_b == out128_b)\n /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ in128x6_ptr\n + 16 `op_Multiply` len128x6 < pow2_64 /\\ out128x6_ptr + 16 `op_Multiply` len128x6 < pow2_64 /\\\n in128_ptr + 16 `op_Multiply` len128 < pow2_64 /\\ out128_ptr + 16 `op_Multiply` len128 < pow2_64\n /\\ inout_ptr + 16 < pow2_64 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 auth_b ==\n va_get_reg64 rRdx va_s0 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 abytes_b == 1\n /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128x6_b ==\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out128x6_b /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 out128_b /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128\n in128x6_b == len128x6 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b ==\n len128 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 inout_b == 1 /\\\n plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32 /\\ va_get_reg64 rR9 va_s0 + 32 <\n pow2_64 /\\ Vale.X64.Memory.buffer_addr #Vale.X64.Memory.vuint128 keys_b (va_get_mem_heaplet 0\n va_s0) + 128 < pow2_64 /\\ len128x6 `op_Modulus` 6 == 0 /\\ (len128x6 > 0 ==> len128x6 >= 6) /\\\n 12 + len128x6 + 6 < pow2_32 /\\ (va_mul_nat len128x6 (128 `op_Division` 8) + va_mul_nat len128\n (128 `op_Division` 8) <= plain_num_bytes /\\ plain_num_bytes < va_mul_nat len128x6 (128\n `op_Division` 8) + va_mul_nat len128 (128 `op_Division` 8) + 128 `op_Division` 8) /\\\n (va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) <= va_get_reg64 rRsi va_s0 /\\\n va_get_reg64 rRsi va_s0 < va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) + 128\n `op_Division` 8) /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rRcx va_s0)\n (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 (Vale.AES.AES_s.aes_encrypt_LE alg key\n (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0))) /\\ (let iv_BE =\n Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in let h_LE =\n Vale.AES.AES_s.aes_encrypt_LE alg key (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0)\n in iv_BE == Vale.AES.GCM_s.compute_iv_BE h_LE iv))))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in Vale.X64.Decls.modifies_buffer128 out128_b\n (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1 va_sM) /\\ Vale.X64.Decls.modifies_buffer128\n iv_b (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2 va_sM) /\\\n Vale.X64.Decls.modifies_buffer128 scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3\n va_sM) /\\ Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0)\n (va_get_mem_heaplet 5 va_sM) /\\ Vale.X64.Decls.modifies_buffer128 out128x6_b\n (va_get_mem_heaplet 6 va_s0) (va_get_mem_heaplet 6 va_sM) /\\ plain_num_bytes < pow2_32 /\\\n va_get_reg64 rRsi va_s0 < pow2_32 /\\ (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0\n (va_get_mem_heaplet 2 va_s0) in let auth_raw_quads = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 7 va_s0) abytes_b) in let auth_bytes =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n auth_raw_quads) 0 (va_get_reg64 rRsi va_s0) in let plain_raw_quads = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_s0) in128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0)\n in128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b) in let plain_bytes =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n plain_raw_quads) 0 plain_num_bytes in let cipher_raw_quads = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM)\n out128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM) inout_b) in let cipher_bytes =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n cipher_raw_quads) 0 plain_num_bytes in l_and (l_and (l_and (FStar.Seq.Base.length\n #Vale.Def.Types_s.nat8 auth_bytes < pow2_32) (FStar.Seq.Base.length #Vale.Def.Types_s.nat8\n plain_bytes < pow2_32)) (cipher_bytes == __proj__Mktuple2__item___1 #(FStar.Seq.Base.seq\n Vale.Def.Types_s.nat8) #bool (Vale.AES.GCM_s.gcm_decrypt_LE alg\n (Vale.Def.Words.Seq_s.seq_nat32_to_seq_nat8_LE key) iv plain_bytes auth_bytes expected_tag)))\n (Vale.Def.Types_s.le_quad32_to_bytes (va_get_xmm 8 va_sM) == Vale.AES.GCM.gcm_decrypt_LE_tag\n alg (Vale.Def.Words.Seq_s.seq_nat32_to_seq_nat8_LE key) iv plain_bytes auth_bytes))) /\\\n va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 6 va_sM (va_update_mem_heaplet\n 5 va_sM (va_update_mem_heaplet 3 va_sM (va_update_mem_heaplet 2 va_sM (va_update_mem_heaplet 1\n va_sM (va_update_xmm 15 va_sM (va_update_xmm 14 va_sM (va_update_xmm 13 va_sM (va_update_xmm 12\n va_sM (va_update_xmm 11 va_sM (va_update_xmm 10 va_sM (va_update_xmm 9 va_sM (va_update_xmm 8\n va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5 va_sM (va_update_xmm 4\n va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1 va_sM (va_update_xmm 0\n va_sM (va_update_reg64 rR15 va_sM (va_update_reg64 rR14 va_sM (va_update_reg64 rR13 va_sM\n (va_update_reg64 rR12 va_sM (va_update_reg64 rR11 va_sM (va_update_reg64 rR10 va_sM\n (va_update_reg64 rR9 va_sM (va_update_reg64 rR8 va_sM (va_update_reg64 rRbp va_sM\n (va_update_reg64 rRsi va_sM (va_update_reg64 rRdi va_sM (va_update_reg64 rRdx va_sM\n (va_update_reg64 rRcx va_sM (va_update_reg64 rRbx va_sM (va_update_reg64 rRax va_sM\n (va_update_ok va_sM (va_update_mem va_sM va_s0)))))))))))))))))))))))))))))))))))))))))\n[@\"opaque_to_smt\"]\nlet va_lemma_Gcm_blocks_wrapped va_b0 va_s0 alg offset auth_b abytes_b in128x6_b out128x6_b in128_b\n out128_b inout_b iv_b iv scratch_b key round_keys keys_b hkeys_b expected_tag =\n let (va_mods:va_mods_t) = [va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5;\n va_Mod_mem_heaplet 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14;\n va_Mod_xmm 13; va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8;\n va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm\n 1; va_Mod_xmm 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12;\n va_Mod_reg64 rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp;\n va_Mod_reg64 rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx;\n va_Mod_reg64 rRax; va_Mod_ok; va_Mod_mem] in\n let va_qc = va_qcode_Gcm_blocks_wrapped va_mods alg offset auth_b abytes_b in128x6_b out128x6_b\n in128_b out128_b inout_b iv_b iv scratch_b key round_keys keys_b hkeys_b expected_tag in\n let (va_sM, va_fM, va_g) = va_wp_sound_code_norm (va_code_Gcm_blocks_wrapped alg offset) va_qc\n va_s0 (fun va_s0 va_sM va_g -> let () = va_g in label va_range1\n \"***** POSTCONDITION NOT MET AT line 588 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\ (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64\n (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0) in let\n (in128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 8) (va_get_stack va_s0) in let (out128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack va_s0) in\n let (len128x6:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 24) (va_get_stack va_s0) in let (in128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack va_s0) in\n let (out128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 40) (va_get_stack va_s0) in let (len128:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack va_s0) in\n let (inout_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 56) (va_get_stack va_s0) in let (plain_num_bytes:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack va_s0) in\n label va_range1\n \"***** POSTCONDITION NOT MET AT line 711 column 56 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out128_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 712 column 52 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 iv_b (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 713 column 57 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 714 column 55 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 715 column 58 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out128x6_b (va_get_mem_heaplet 6 va_s0) (va_get_mem_heaplet\n 6 va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 718 column 39 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (plain_num_bytes < pow2_32) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 719 column 38 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRsi va_s0 < pow2_32) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 721 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 723 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let auth_raw_quads = FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) auth_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 7 va_s0)\n abytes_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 724 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let auth_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes auth_raw_quads) 0 (va_get_reg64 rRsi va_s0) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 725 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let plain_raw_quads = FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 726 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let plain_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes plain_raw_quads) 0 plain_num_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 727 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let cipher_raw_quads = FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM) out128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 728 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let cipher_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes cipher_raw_quads) 0 plain_num_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 737 column 52 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (l_and (l_and (l_and (FStar.Seq.Base.length #Vale.Def.Types_s.nat8 auth_bytes < pow2_32)\n (FStar.Seq.Base.length #Vale.Def.Types_s.nat8 plain_bytes < pow2_32)) (cipher_bytes ==\n __proj__Mktuple2__item___1 #(FStar.Seq.Base.seq Vale.Def.Types_s.nat8) #bool\n (Vale.AES.GCM_s.gcm_decrypt_LE alg (Vale.Def.Words.Seq_s.seq_nat32_to_seq_nat8_LE key) iv\n plain_bytes auth_bytes expected_tag))) (Vale.Def.Types_s.le_quad32_to_bytes (va_get_xmm 8\n va_sM) == Vale.AES.GCM.gcm_decrypt_LE_tag alg (Vale.Def.Words.Seq_s.seq_nat32_to_seq_nat8_LE\n key) iv plain_bytes auth_bytes))))))))))) in\n assert_norm (va_qc.mods == va_mods);\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet\n 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13;\n va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm\n 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64\n rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64\n rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64\n rRax; va_Mod_ok; va_Mod_mem]) va_sM va_s0;\n (va_sM, va_fM)\n\n[@ va_qattr]\nlet va_wp_Gcm_blocks_wrapped (alg:algorithm) (offset:int) (auth_b:buffer128) (abytes_b:buffer128)\n (in128x6_b:buffer128) (out128x6_b:buffer128) (in128_b:buffer128) (out128_b:buffer128)\n (inout_b:buffer128) (iv_b:buffer128) (iv:supported_iv_LE) (scratch_b:buffer128) (key:(seq nat32))\n (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (expected_tag:(seq nat8))\n (va_s0:va_state) (va_k:(va_state -> unit -> Type0)) : Type0 =\n (va_get_ok va_s0 /\\ (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64\n (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0) in let\n (in128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 8) (va_get_stack va_s0) in let (out128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack va_s0) in\n let (len128x6:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 24) (va_get_stack va_s0) in let (in128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack va_s0) in\n let (out128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 40) (va_get_stack va_s0) in let (len128:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack va_s0) in\n let (inout_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 56) (va_get_stack va_s0) in let (plain_num_bytes:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack va_s0) in\n sse_enabled /\\ movbe_enabled /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 +\n offset + 0) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0)\n Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 32) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 48) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 64) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0) (va_get_reg64 rRdi va_s0) auth_b\n (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 7 va_s0) abytes_ptr abytes_b 1 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 2 va_s0) (va_get_reg64 rR8 va_s0) iv_b 1\n (va_get_mem_layout va_s0) Public /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 6\n va_s0) in128x6_ptr in128x6_b len128x6 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 6 va_s0) out128x6_ptr out128x6_b len128x6\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1\n va_s0) in128_ptr in128_b len128 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1 va_s0) out128_ptr out128_b len128\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5\n va_s0) inout_ptr inout_b 1 (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128\n (va_get_mem_heaplet 3 va_s0) (va_get_reg64 rRbp va_s0) scratch_b 9 (va_get_mem_layout va_s0)\n Secret /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0)\n hkeys_b 8 (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.buffer_disjoints128 iv_b ([keys_b;\n scratch_b; in128x6_b; out128x6_b; hkeys_b; in128_b; out128_b; inout_b]) /\\\n Vale.X64.Decls.buffer_disjoints128 scratch_b ([keys_b; in128x6_b; out128x6_b; in128_b;\n out128_b; inout_b; hkeys_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128x6_b ([keys_b;\n hkeys_b; in128_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128_b ([keys_b; hkeys_b;\n out128x6_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 inout_b ([keys_b; hkeys_b;\n out128x6_b; out128_b]) /\\ (Vale.X64.Decls.buffers_disjoint128 in128x6_b out128x6_b \\/ in128x6_b\n == out128x6_b) /\\ (Vale.X64.Decls.buffers_disjoint128 in128_b out128_b \\/ in128_b == out128_b)\n /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ in128x6_ptr\n + 16 `op_Multiply` len128x6 < pow2_64 /\\ out128x6_ptr + 16 `op_Multiply` len128x6 < pow2_64 /\\\n in128_ptr + 16 `op_Multiply` len128 < pow2_64 /\\ out128_ptr + 16 `op_Multiply` len128 < pow2_64\n /\\ inout_ptr + 16 < pow2_64 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 auth_b ==\n va_get_reg64 rRdx va_s0 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 abytes_b == 1\n /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128x6_b ==\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out128x6_b /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 out128_b /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128\n in128x6_b == len128x6 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b ==\n len128 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 inout_b == 1 /\\\n plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32 /\\ va_get_reg64 rR9 va_s0 + 32 <\n pow2_64 /\\ Vale.X64.Memory.buffer_addr #Vale.X64.Memory.vuint128 keys_b (va_get_mem_heaplet 0\n va_s0) + 128 < pow2_64 /\\ len128x6 `op_Modulus` 6 == 0 /\\ (len128x6 > 0 ==> len128x6 >= 6) /\\\n 12 + len128x6 + 6 < pow2_32 /\\ (va_mul_nat len128x6 (128 `op_Division` 8) + va_mul_nat len128\n (128 `op_Division` 8) <= plain_num_bytes /\\ plain_num_bytes < va_mul_nat len128x6 (128\n `op_Division` 8) + va_mul_nat len128 (128 `op_Division` 8) + 128 `op_Division` 8) /\\\n (va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) <= va_get_reg64 rRsi va_s0 /\\\n va_get_reg64 rRsi va_s0 < va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) + 128\n `op_Division` 8) /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rRcx va_s0)\n (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 (Vale.AES.AES_s.aes_encrypt_LE alg key\n (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0))) /\\ (let iv_BE =\n Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in let h_LE =\n Vale.AES.AES_s.aes_encrypt_LE alg key (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0)\n in iv_BE == Vale.AES.GCM_s.compute_iv_BE h_LE iv)) /\\ (forall (va_x_mem:vale_heap)\n (va_x_rax:nat64) (va_x_rbx:nat64) (va_x_rcx:nat64) (va_x_rdx:nat64) (va_x_rdi:nat64)\n (va_x_rsi:nat64) (va_x_rbp:nat64) (va_x_r8:nat64) (va_x_r9:nat64) (va_x_r10:nat64)\n (va_x_r11:nat64) (va_x_r12:nat64) (va_x_r13:nat64) (va_x_r14:nat64) (va_x_r15:nat64)\n (va_x_xmm0:quad32) (va_x_xmm1:quad32) (va_x_xmm2:quad32) (va_x_xmm3:quad32) (va_x_xmm4:quad32)\n (va_x_xmm5:quad32) (va_x_xmm6:quad32) (va_x_xmm7:quad32) (va_x_xmm8:quad32) (va_x_xmm9:quad32)\n (va_x_xmm10:quad32) (va_x_xmm11:quad32) (va_x_xmm12:quad32) (va_x_xmm13:quad32)\n (va_x_xmm14:quad32) (va_x_xmm15:quad32) (va_x_heap1:vale_heap) (va_x_heap2:vale_heap)\n (va_x_heap3:vale_heap) (va_x_heap5:vale_heap) (va_x_heap6:vale_heap)\n (va_x_efl:Vale.X64.Flags.t) . let va_sM = va_upd_flags va_x_efl (va_upd_mem_heaplet 6\n va_x_heap6 (va_upd_mem_heaplet 5 va_x_heap5 (va_upd_mem_heaplet 3 va_x_heap3\n (va_upd_mem_heaplet 2 va_x_heap2 (va_upd_mem_heaplet 1 va_x_heap1 (va_upd_xmm 15 va_x_xmm15\n (va_upd_xmm 14 va_x_xmm14 (va_upd_xmm 13 va_x_xmm13 (va_upd_xmm 12 va_x_xmm12 (va_upd_xmm 11\n va_x_xmm11 (va_upd_xmm 10 va_x_xmm10 (va_upd_xmm 9 va_x_xmm9 (va_upd_xmm 8 va_x_xmm8\n (va_upd_xmm 7 va_x_xmm7 (va_upd_xmm 6 va_x_xmm6 (va_upd_xmm 5 va_x_xmm5 (va_upd_xmm 4 va_x_xmm4\n (va_upd_xmm 3 va_x_xmm3 (va_upd_xmm 2 va_x_xmm2 (va_upd_xmm 1 va_x_xmm1 (va_upd_xmm 0 va_x_xmm0\n (va_upd_reg64 rR15 va_x_r15 (va_upd_reg64 rR14 va_x_r14 (va_upd_reg64 rR13 va_x_r13\n (va_upd_reg64 rR12 va_x_r12 (va_upd_reg64 rR11 va_x_r11 (va_upd_reg64 rR10 va_x_r10\n (va_upd_reg64 rR9 va_x_r9 (va_upd_reg64 rR8 va_x_r8 (va_upd_reg64 rRbp va_x_rbp (va_upd_reg64\n rRsi va_x_rsi (va_upd_reg64 rRdi va_x_rdi (va_upd_reg64 rRdx va_x_rdx (va_upd_reg64 rRcx\n va_x_rcx (va_upd_reg64 rRbx va_x_rbx (va_upd_reg64 rRax va_x_rax (va_upd_mem va_x_mem\n va_s0))))))))))))))))))))))))))))))))))))) in va_get_ok va_sM /\\ (let\n (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in Vale.X64.Decls.modifies_buffer128 out128_b\n (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1 va_sM) /\\ Vale.X64.Decls.modifies_buffer128\n iv_b (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2 va_sM) /\\\n Vale.X64.Decls.modifies_buffer128 scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3\n va_sM) /\\ Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0)\n (va_get_mem_heaplet 5 va_sM) /\\ Vale.X64.Decls.modifies_buffer128 out128x6_b\n (va_get_mem_heaplet 6 va_s0) (va_get_mem_heaplet 6 va_sM) /\\ plain_num_bytes < pow2_32 /\\\n va_get_reg64 rRsi va_s0 < pow2_32 /\\ (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0\n (va_get_mem_heaplet 2 va_s0) in let auth_raw_quads = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 7 va_s0) abytes_b) in let auth_bytes =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n auth_raw_quads) 0 (va_get_reg64 rRsi va_s0) in let plain_raw_quads = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_s0) in128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0)\n in128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b) in let plain_bytes =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n plain_raw_quads) 0 plain_num_bytes in let cipher_raw_quads = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM)\n out128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM) inout_b) in let cipher_bytes =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n cipher_raw_quads) 0 plain_num_bytes in l_and (l_and (l_and (FStar.Seq.Base.length\n #Vale.Def.Types_s.nat8 auth_bytes < pow2_32) (FStar.Seq.Base.length #Vale.Def.Types_s.nat8\n plain_bytes < pow2_32)) (cipher_bytes == __proj__Mktuple2__item___1 #(FStar.Seq.Base.seq\n Vale.Def.Types_s.nat8) #bool (Vale.AES.GCM_s.gcm_decrypt_LE alg\n (Vale.Def.Words.Seq_s.seq_nat32_to_seq_nat8_LE key) iv plain_bytes auth_bytes expected_tag)))\n (Vale.Def.Types_s.le_quad32_to_bytes (va_get_xmm 8 va_sM) == Vale.AES.GCM.gcm_decrypt_LE_tag\n alg (Vale.Def.Words.Seq_s.seq_nat32_to_seq_nat8_LE key) iv plain_bytes auth_bytes))) ==> va_k\n va_sM (())))\n\n\nval va_wpProof_Gcm_blocks_wrapped : alg:algorithm -> offset:int -> auth_b:buffer128 ->\n abytes_b:buffer128 -> in128x6_b:buffer128 -> out128x6_b:buffer128 -> in128_b:buffer128 ->\n out128_b:buffer128 -> inout_b:buffer128 -> iv_b:buffer128 -> iv:supported_iv_LE ->\n scratch_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128 -> expected_tag:(seq nat8) -> va_s0:va_state -> va_k:(va_state -> unit -> Type0)\n -> Ghost (va_state & va_fuel & unit)\n (requires (va_t_require va_s0 /\\ va_wp_Gcm_blocks_wrapped alg offset auth_b abytes_b in128x6_b\n out128x6_b in128_b out128_b inout_b iv_b iv scratch_b key round_keys keys_b hkeys_b\n expected_tag va_s0 va_k))\n (ensures (fun (va_sM, va_f0, va_g) -> va_t_ensure (va_code_Gcm_blocks_wrapped alg offset)\n ([va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet 3;\n va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13;\n va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm\n 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64\n rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64\n rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64\n rRax; va_Mod_mem]) va_s0 va_k ((va_sM, va_f0, va_g))))\n[@\"opaque_to_smt\"]\nlet va_wpProof_Gcm_blocks_wrapped alg offset auth_b abytes_b in128x6_b out128x6_b in128_b out128_b\n inout_b iv_b iv scratch_b key round_keys keys_b hkeys_b expected_tag va_s0 va_k =\n let (va_sM, va_f0) = va_lemma_Gcm_blocks_wrapped (va_code_Gcm_blocks_wrapped alg offset) va_s0\n alg offset auth_b abytes_b in128x6_b out128x6_b in128_b out128_b inout_b iv_b iv scratch_b key\n round_keys keys_b hkeys_b expected_tag in\n va_lemma_upd_update va_sM;\n assert (va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 6 va_sM\n (va_update_mem_heaplet 5 va_sM (va_update_mem_heaplet 3 va_sM (va_update_mem_heaplet 2 va_sM\n (va_update_mem_heaplet 1 va_sM (va_update_xmm 15 va_sM (va_update_xmm 14 va_sM (va_update_xmm\n 13 va_sM (va_update_xmm 12 va_sM (va_update_xmm 11 va_sM (va_update_xmm 10 va_sM (va_update_xmm\n 9 va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5\n va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1\n va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR15 va_sM (va_update_reg64 rR14 va_sM\n (va_update_reg64 rR13 va_sM (va_update_reg64 rR12 va_sM (va_update_reg64 rR11 va_sM\n (va_update_reg64 rR10 va_sM (va_update_reg64 rR9 va_sM (va_update_reg64 rR8 va_sM\n (va_update_reg64 rRbp va_sM (va_update_reg64 rRsi va_sM (va_update_reg64 rRdi va_sM\n (va_update_reg64 rRdx va_sM (va_update_reg64 rRcx va_sM (va_update_reg64 rRbx va_sM\n (va_update_reg64 rRax va_sM (va_update_ok va_sM (va_update_mem va_sM\n va_s0))))))))))))))))))))))))))))))))))))))));\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet\n 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13;\n va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm\n 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64\n rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64\n rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64\n rRax; va_Mod_mem]) va_sM va_s0;\n let va_g = () in\n (va_sM, va_f0, va_g)\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_quick_Gcm_blocks_wrapped (alg:algorithm) (offset:int) (auth_b:buffer128)\n (abytes_b:buffer128) (in128x6_b:buffer128) (out128x6_b:buffer128) (in128_b:buffer128)\n (out128_b:buffer128) (inout_b:buffer128) (iv_b:buffer128) (iv:supported_iv_LE)\n (scratch_b:buffer128) (key:(seq nat32)) (round_keys:(seq quad32)) (keys_b:buffer128)\n (hkeys_b:buffer128) (expected_tag:(seq nat8)) : (va_quickCode unit (va_code_Gcm_blocks_wrapped\n alg offset)) =\n (va_QProc (va_code_Gcm_blocks_wrapped alg offset) ([va_Mod_flags; va_Mod_mem_heaplet 6;\n va_Mod_mem_heaplet 5; va_Mod_mem_heaplet 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1;\n va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13; va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10;\n va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm\n 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64\n rR13; va_Mod_reg64 rR12; va_Mod_reg64 rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64\n rR8; va_Mod_reg64 rRbp; va_Mod_reg64 rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64\n rRcx; va_Mod_reg64 rRbx; va_Mod_reg64 rRax; va_Mod_mem]) (va_wp_Gcm_blocks_wrapped alg offset\n auth_b abytes_b in128x6_b out128x6_b in128_b out128_b inout_b iv_b iv scratch_b key round_keys\n keys_b hkeys_b expected_tag) (va_wpProof_Gcm_blocks_wrapped alg offset auth_b abytes_b\n in128x6_b out128x6_b in128_b out128_b inout_b iv_b iv scratch_b key round_keys keys_b hkeys_b\n expected_tag))\n#pop-options\n//--\n//-- Gcm_blocks_decrypt_stdcall\n#push-options \"--z3rlimit 1600\"\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_blocks_decrypt_stdcall win alg =\n (va_Block (va_CCons (va_code_CreateHeaplets ()) (va_CCons (va_Block (va_CNil ())) (va_CCons\n (va_code_Save_registers win) (va_CCons (if win then va_Block (va_CCons (va_code_Mov64\n (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRcx)) (va_CCons (va_code_Mov64\n (va_op_dst_opr64_reg64 rRsi) (va_op_opr64_reg64 rRdx)) (va_CCons (va_code_Mov64\n (va_op_dst_opr64_reg64 rRdx) (va_op_opr64_reg64 rR8)) (va_CCons (va_code_Mov64\n (va_op_dst_opr64_reg64 rRcx) (va_op_opr64_reg64 rR9)) (va_CCons (va_code_Load64_stack\n (va_op_dst_opr64_reg64 rR8) (va_op_reg_opr64_reg64 rRsp) (224 + 40 + 0)) (va_CCons\n (va_code_Load64_stack (va_op_dst_opr64_reg64 rR9) (va_op_reg_opr64_reg64 rRsp) (224 + 40 + 8))\n (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rRbp) (va_op_reg_opr64_reg64 rRsp) (224\n + 40 + 88)) (va_CNil ())))))))) else va_Block (va_CCons (va_code_Load64_stack\n (va_op_dst_opr64_reg64 rRbp) (va_op_reg_opr64_reg64 rRsp) (64 + 8 + 72)) (va_CNil ())))\n (va_CCons (va_code_Gcm_blocks_wrapped alg (Vale.X64.Decls.total_if #int win (224 + 56) (64 +\n 8))) (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rR15) (va_op_reg_opr64_reg64 rRsp)\n (Vale.X64.Decls.total_if #int win (224 + 40 + 96) (64 + 8 + 80))) (va_CCons\n (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 0) (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64\n rR15) 0 Secret) (va_CCons (va_code_XmmEqual (va_op_xmm_xmm 0) (va_op_xmm_xmm 8)) (va_CCons\n (va_Block (va_CNil ())) (va_CCons (va_Block (va_CNil ())) (va_CCons (va_code_Mov64\n (va_op_dst_opr64_reg64 rRcx) (va_op_opr64_reg64 rRax)) (va_CCons (va_code_Restore_registers\n win) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRax) (va_op_opr64_reg64 rRcx)) (va_CCons\n (va_code_DestroyHeaplets ()) (va_CNil ()))))))))))))))))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_blocks_decrypt_stdcall win alg =\n (va_pbool_and (va_codegen_success_CreateHeaplets ()) (va_pbool_and\n (va_codegen_success_Save_registers win) (va_pbool_and (if win then va_pbool_and\n (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRcx)) (va_pbool_and\n (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rRsi) (va_op_opr64_reg64 rRdx)) (va_pbool_and\n (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rRdx) (va_op_opr64_reg64 rR8)) (va_pbool_and\n (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rRcx) (va_op_opr64_reg64 rR9)) (va_pbool_and\n (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rR8) (va_op_reg_opr64_reg64 rRsp) (224\n + 40 + 0)) (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rR9)\n (va_op_reg_opr64_reg64 rRsp) (224 + 40 + 8)) (va_pbool_and (va_codegen_success_Load64_stack\n (va_op_dst_opr64_reg64 rRbp) (va_op_reg_opr64_reg64 rRsp) (224 + 40 + 88)) (va_ttrue ())))))))\n else va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRbp)\n (va_op_reg_opr64_reg64 rRsp) (64 + 8 + 72)) (va_ttrue ())) (va_pbool_and\n (va_codegen_success_Gcm_blocks_wrapped alg (Vale.X64.Decls.total_if #int win (224 + 56) (64 +\n 8))) (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rR15)\n (va_op_reg_opr64_reg64 rRsp) (Vale.X64.Decls.total_if #int win (224 + 40 + 96) (64 + 8 + 80)))\n (va_pbool_and (va_codegen_success_Load128_buffer (va_op_heaplet_mem_heaplet 0) (va_op_xmm_xmm\n 0) (va_op_reg_opr64_reg64 rR15) 0 Secret) (va_pbool_and (va_codegen_success_XmmEqual\n (va_op_xmm_xmm 0) (va_op_xmm_xmm 8)) (va_pbool_and (va_codegen_success_Mov64\n (va_op_dst_opr64_reg64 rRcx) (va_op_opr64_reg64 rRax)) (va_pbool_and\n (va_codegen_success_Restore_registers win) (va_pbool_and (va_codegen_success_Mov64\n (va_op_dst_opr64_reg64 rRax) (va_op_opr64_reg64 rRcx)) (va_pbool_and\n (va_codegen_success_DestroyHeaplets ()) (va_ttrue ()))))))))))))\n\nPartial implementation:\nlet va_qcode_Gcm_blocks_decrypt_stdcall\n (va_mods: va_mods_t)\n (win: bool)\n (alg: algorithm)\n (auth_b: buffer128)\n (auth_bytes auth_num: nat64)\n (keys_b iv_b: buffer128)\n (iv: supported_iv_LE)\n (hkeys_b abytes_b in128x6_b out128x6_b: buffer128)\n (len128x6_num: nat64)\n (in128_b out128_b: buffer128)\n (len128_num: nat64)\n (inout_b: buffer128)\n (cipher_num: nat64)\n (scratch_b tag_b: buffer128)\n (key: (seq nat32))\n : (va_quickCode unit (va_code_Gcm_blocks_decrypt_stdcall win alg)) =\n \n"
},
{
"role": "assistant",
"content": "(qblock va_mods\n (fun (va_s: va_state) ->\n let va_old_s:va_state = va_s in\n let auth_ptr:(va_int_range 0 18446744073709551615) =\n va_if win (fun _ -> va_get_reg64 rRcx va_s) (fun _ -> va_get_reg64 rRdi va_s)\n in\n let auth_num_bytes:(va_int_range 0 18446744073709551615) =\n va_if win (fun _ -> va_get_reg64 rRdx va_s) (fun _ -> va_get_reg64 rRsi va_s)\n in\n let auth_len:(va_int_range 0 18446744073709551615) =\n va_if win (fun _ -> va_get_reg64 rR8 va_s) (fun _ -> va_get_reg64 rRdx va_s)\n in\n let keys_ptr:(va_int_range 0 18446744073709551615) =\n va_if win (fun _ -> va_get_reg64 rR9 va_s) (fun _ -> va_get_reg64 rRcx va_s)\n in\n let iv_ptr:(va_int_range 0 18446744073709551615) =\n va_if win\n (fun _ ->\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + 32 + 8 + 0)\n (va_get_stack va_s))\n (fun _ -> va_get_reg64 rR8 va_s)\n in\n let xip:(va_int_range 0 18446744073709551615) =\n va_if win\n (fun _ ->\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + 32 + 8 + 8)\n (va_get_stack va_s))\n (fun _ -> va_get_reg64 rR9 va_s)\n in\n let abytes_ptr:(va_int_range 0 18446744073709551615) =\n va_if win\n (fun _ ->\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + 40 + 16) (va_get_stack va_s)\n )\n (fun _ ->\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + 8 + 0) (va_get_stack va_s))\n in\n let in128x6_ptr:(va_int_range 0 18446744073709551615) =\n va_if win\n (fun _ ->\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + 40 + 24) (va_get_stack va_s)\n )\n (fun _ ->\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + 8 + 8) (va_get_stack va_s))\n in\n let out128x6_ptr:(va_int_range 0 18446744073709551615) =\n va_if win\n (fun _ ->\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + 40 + 32) (va_get_stack va_s)\n )\n (fun _ ->\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + 8 + 16) (va_get_stack va_s))\n in\n let len128x6:(va_int_range 0 18446744073709551615) =\n va_if win\n (fun _ ->\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + 40 + 40) (va_get_stack va_s)\n )\n (fun _ ->\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + 8 + 24) (va_get_stack va_s))\n in\n let in128_ptr:(va_int_range 0 18446744073709551615) =\n va_if win\n (fun _ ->\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + 40 + 48) (va_get_stack va_s)\n )\n (fun _ ->\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + 8 + 32) (va_get_stack va_s))\n in\n let out128_ptr:(va_int_range 0 18446744073709551615) =\n va_if win\n (fun _ ->\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + 40 + 56) (va_get_stack va_s)\n )\n (fun _ ->\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + 8 + 40) (va_get_stack va_s))\n in\n let len128:(va_int_range 0 18446744073709551615) =\n va_if win\n (fun _ ->\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + 40 + 64) (va_get_stack va_s)\n )\n (fun _ ->\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + 8 + 48) (va_get_stack va_s))\n in\n let inout_ptr:(va_int_range 0 18446744073709551615) =\n va_if win\n (fun _ ->\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + 40 + 72) (va_get_stack va_s)\n )\n (fun _ ->\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + 8 + 56) (va_get_stack va_s))\n in\n let cipher_num_bytes:(va_int_range 0 18446744073709551615) =\n va_if win\n (fun _ ->\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + 40 + 80) (va_get_stack va_s)\n )\n (fun _ ->\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + 8 + 64) (va_get_stack va_s))\n in\n let scratch_ptr:(va_int_range 0 18446744073709551615) =\n va_if win\n (fun _ ->\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + 40 + 88) (va_get_stack va_s)\n )\n (fun _ ->\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + 8 + 72) (va_get_stack va_s))\n in\n let tag_ptr:(va_int_range 0 18446744073709551615) =\n va_if win\n (fun _ ->\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + 40 + 96) (va_get_stack va_s)\n )\n (fun _ ->\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + 8 + 80) (va_get_stack va_s))\n in\n va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 989 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_CreateHeaplets ([\n declare_buffer128 auth_b 1 Secret Immutable;\n declare_buffer128 abytes_b 7 Secret Immutable;\n declare_buffer128 in128x6_b 6 Secret Immutable;\n declare_buffer128 in128_b 1 Secret Immutable;\n declare_buffer128 hkeys_b 0 Secret Immutable;\n declare_buffer128 tag_b 0 Secret Immutable;\n declare_buffer128 keys_b 0 Secret Immutable;\n declare_buffer128 iv_b 2 Public Mutable;\n declare_buffer128 scratch_b 3 Secret Mutable;\n declare_buffer128 out128x6_b 6 Secret Mutable;\n declare_buffer128 out128_b 1 Secret Mutable;\n declare_buffer128 inout_b 5 Secret Mutable\n ]))\n (fun (va_s: va_state) _ ->\n va_qAssertSquash va_range1\n \"***** EXPRESSION PRECONDITIONS NOT MET WITHIN line 1003 column 30 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n ((fun\n (alg_10639: Vale.AES.AES_common_s.algorithm)\n (key_10640: (FStar.Seq.Base.seq Vale.Def.Types_s.nat32))\n (input_LE_10641: Vale.Def.Types_s.quad32)\n ->\n Vale.AES.AES_s.is_aes_key_LE alg_10639 key_10640) alg\n key\n (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0))\n (fun _ ->\n let va_arg69:Vale.Def.Types_s.quad32 =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.AES.AES_s.aes_encrypt_LE alg\n key\n (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0))\n in\n let va_arg68:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32) =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s) hkeys_b\n in\n va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 1003 column 30 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_: unit) -> Vale.AES.GHash.lemma_hkeys_reqs_pub_priv va_arg68 va_arg69)\n (va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 1004 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (win ==>\n Vale.X64.Stack_i.valid_src_stack64 (va_get_reg64 rRsp va_s + 40)\n (va_get_stack va_s))\n (va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 1005 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (~win ==>\n Vale.X64.Stack_i.valid_src_stack64 (va_get_reg64 rRsp va_s + 80)\n (va_get_stack va_s))\n (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 1006 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Save_registers win)\n (fun (va_s: va_state) _ ->\n va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 1009 column 15 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_qInlineIf va_mods\n win\n (qblock va_mods\n (fun (va_s: va_state) ->\n va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 1011 column 14 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRdi)\n (va_op_opr64_reg64 rRcx))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 1012 column 14 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRsi\n )\n (va_op_opr64_reg64 rRdx))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 1013 column 14 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64\n rRdx)\n (va_op_opr64_reg64 rR8))\n (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 1014 column 14 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64\n rRcx)\n (va_op_opr64_reg64 rR9))\n (fun (va_s: va_state) _ ->\n va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 1015 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRsp va_s +\n 224 ==\n va_get_reg64 rRsp va_old_s\n )\n (va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 1016 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Stack_i.valid_src_stack64\n (va_get_reg64 rRsp\n va_s +\n 224 +\n 40)\n (va_get_stack va_s))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 1017 column 21 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack\n (va_op_dst_opr64_reg64\n rR8)\n (va_op_reg_opr64_reg64\n rRsp)\n (224 + 40 + 0))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 1018 column 21 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack\n (va_op_dst_opr64_reg64\n rR9)\n (va_op_reg_opr64_reg64\n rRsp)\n (224 + 40 +\n 8))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 1019 column 21 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack\n (va_op_dst_opr64_reg64\n rRbp\n )\n (va_op_reg_opr64_reg64\n rRsp\n )\n (224 +\n 40 +\n 88))\n (va_QEmpty (()\n ))))))))\n ))))\n (qblock va_mods\n (fun (va_s: va_state) ->\n va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 1023 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRsp va_s + 64 ==\n va_get_reg64 rRsp va_old_s)\n (va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 1024 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Stack_i.valid_src_stack64 (va_get_reg64\n rRsp\n va_s +\n 64 +\n 80)\n (va_get_stack va_s))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 1025 column 21 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64\n rRbp)\n (va_op_reg_opr64_reg64 rRsp)\n (64 + 8 + 72))\n (va_QEmpty (())))))))\n (fun (va_s: va_state) va_g ->\n va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 1028 column 23 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_blocks_wrapped alg\n (Vale.X64.Decls.total_if #int\n win\n (224 + 56)\n (64 + 8)) auth_b abytes_b in128x6_b out128x6_b\n in128_b out128_b inout_b iv_b iv scratch_b key\n (Vale.X64.Decls.buffer128_as_seq (va_get_mem_heaplet\n 0\n va_old_s)\n keys_b) keys_b hkeys_b\n (Vale.Def.Types_s.le_quad32_to_bytes (Vale.X64.Decls.buffer128_read\n tag_b\n 0\n (va_get_mem_heaplet 0 va_old_s))))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 1048 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rR15\n )\n (va_op_reg_opr64_reg64 rRsp)\n (Vale.X64.Decls.total_if #int\n win\n (224 + 40 + 96)\n (64 + 8 + 80)))\n (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 1049 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet\n 0)\n (va_op_xmm_xmm 0)\n (va_op_reg_opr64_reg64 rR15)\n 0\n Secret\n tag_b\n 0)\n (fun (va_s: va_state) _ ->\n va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 1050 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_xmm 0 va_s ==\n Vale.X64.Decls.buffer128_read tag_b\n 0\n (va_get_mem_heaplet 0 va_s))\n (va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 1051 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_xmm 0 va_s ==\n Vale.X64.Decls.buffer128_read tag_b\n 0\n (va_get_mem_heaplet 0 va_old_s))\n (let alleged_tag_quad:quad32 =\n va_get_xmm 0 va_s\n in\n let computed_tag:quad32 =\n va_get_xmm 8 va_s\n in\n va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 1054 column 13 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_XmmEqual (va_op_xmm_xmm\n 0)\n (va_op_xmm_xmm 8))\n (fun (va_s: va_state) _ ->\n let auth_raw_quads:(FStar.Seq.Base.seq\n Vale.X64.Decls.quad32) =\n FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet\n 1\n va_old_s)\n auth_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet\n 7\n va_old_s)\n abytes_b)\n in\n va_qAssertSquash va_range1\n \"***** EXPRESSION PRECONDITIONS NOT MET WITHIN line 1057 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n ((fun\n a_1906\n (s_1907:\n (FStar.Seq.Base.seq\n a_1906))\n (i_1908: Prims.nat)\n (j_1909: Prims.nat)\n ->\n let j_1869:Prims.nat\n =\n j_1909\n in\n Prims.b2t (Prims.op_AmpAmp\n (Prims.op_LessThanOrEqual\n i_1908\n j_1869)\n (Prims.op_LessThanOrEqual\n j_1869\n (FStar.Seq.Base.length\n #a_1906\n s_1907\n )))) Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes\n auth_raw_quads)\n 0\n auth_num_bytes)\n (fun _ ->\n let auth_bytes:(FStar.Seq.Base.seq\n Vale.Def.Types_s.nat8\n ) =\n FStar.Seq.Base.slice\n #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes\n auth_raw_quads\n )\n 0\n auth_num_bytes\n in\n let cipher_raw_quads:(FStar.Seq.Base.seq\n Vale.X64.Decls.quad32\n ) =\n FStar.Seq.Base.append\n #Vale.X64.Decls.quad32\n (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128\n (va_get_mem_heaplet\n 6\n va_old_s\n )\n in128x6_b)\n (Vale.X64.Decls.s128\n (va_get_mem_heaplet\n 1\n va_old_s\n )\n in128_b))\n (Vale.X64.Decls.s128\n (va_get_mem_heaplet\n 5\n va_old_s)\n inout_b)\n in\n va_qAssertSquash va_range1\n \"***** EXPRESSION PRECONDITIONS NOT MET WITHIN line 1059 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n ((fun\n a_1906\n (s_1907:\n (FStar.Seq.Base.seq\n a_1906))\n (i_1908:\n Prims.nat)\n (j_1909:\n Prims.nat)\n ->\n let j_1869:Prims.nat\n =\n j_1909\n in\n Prims.b2t (Prims.op_AmpAmp\n (Prims.op_LessThanOrEqual\n i_1908\n j_1869\n )\n (Prims.op_LessThanOrEqual\n j_1869\n (FStar.Seq.Base.length\n #a_1906\n s_1907\n ))\n )) Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes\n cipher_raw_quads\n )\n 0\n cipher_num_bytes\n )\n (fun _ ->\n let cipher_bytes:(\n FStar.Seq.Base.seq\n Vale.Def.Types_s.nat8\n ) =\n FStar.Seq.Base.slice\n #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes\n cipher_raw_quads\n )\n 0\n cipher_num_bytes\n\n in\n let va_arg67:Vale.Def.Types_s.quad32\n =\n computed_tag\n in\n let va_arg66:Vale.Def.Types_s.quad32\n =\n alleged_tag_quad\n in\n let va_arg65:Vale.Def.Words_s.nat64\n =\n va_get_reg64 rRax\n va_s\n in\n let va_arg64:(FStar.Seq.Base.seq\n Vale.Def.Words_s.nat8\n ) =\n auth_bytes\n in\n let va_arg63:(FStar.Seq.Base.seq\n Vale.Def.Words_s.nat8\n ) =\n cipher_bytes\n in\n let va_arg62:Vale.AES.GCM_s.supported_iv_LE\n =\n iv\n in\n let va_arg61:(FStar.Seq.Base.seq\n Vale.Def.Words_s.nat8\n ) =\n Vale.Def.Words.Seq_s.seq_nat32_to_seq_nat8_LE\n key\n in\n let va_arg60:Vale.AES.AES_common_s.algorithm\n =\n alg\n in\n va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 1060 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun\n (_: unit)\n ->\n Vale.AES.GCM.decrypt_helper\n va_arg60\n va_arg61\n va_arg62\n va_arg63\n va_arg64\n va_arg65\n va_arg66\n va_arg67\n )\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 1066 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64\n (va_op_dst_opr64_reg64\n rRcx\n )\n (va_op_opr64_reg64\n rRax\n ))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 1067 column 22 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Restore_registers\n win\n (va_get_reg64\n rRsp\n va_old_s\n\n )\n (va_get_xmm\n 6\n va_old_s\n\n )\n (va_get_xmm\n 7\n va_old_s\n\n )\n (va_get_xmm\n 8\n va_old_s\n\n )\n (va_get_xmm\n 9\n va_old_s\n\n )\n (va_get_xmm\n 10\n va_old_s\n\n )\n (va_get_xmm\n 11\n va_old_s\n\n )\n (va_get_xmm\n 12\n va_old_s\n\n )\n (va_get_xmm\n 13\n va_old_s\n\n )\n (va_get_xmm\n 14\n va_old_s\n\n )\n (va_get_xmm\n 15\n va_old_s\n\n ))\n (va_QSeq\n va_range1\n \"***** PRECONDITION NOT MET AT line 1068 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64\n (\n va_op_dst_opr64_reg64\n rRax\n\n )\n (\n va_op_opr64_reg64\n rRcx\n\n )\n\n )\n (va_QSeq\n va_range1\n \"***** PRECONDITION NOT MET AT line 1070 column 20 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (\n va_quick_DestroyHeaplets\n ()\n\n )\n (\n va_QEmpty\n (\n ()\n )\n\n )\n\n ))\n ))))))))))\n )))))))))"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: t: Hacl.Bignum.Definitions.limb_t -> len: Hacl.Bignum.meta_len t -> Type0\nEffect: Prims.Tot\n\nContext:\nmodule Hacl.Bignum.ModInv\n\nopen FStar.HyperStack\nopen FStar.HyperStack.ST\nopen FStar.Mul\n\nopen Lib.IntTypes\nopen Lib.Buffer\n\nopen Hacl.Bignum.Definitions\n\nmodule ST = FStar.HyperStack.ST\nmodule S = Hacl.Spec.Bignum.ModInv\n\nmodule BN = Hacl.Bignum\nmodule BE = Hacl.Bignum.Exponentiation\nmodule BM = Hacl.Bignum.Montgomery\nmodule SD = Hacl.Spec.Bignum.Definitions\n\n#reset-options \"--z3rlimit 50 --fuel 0 --ifuel 0\"\n\nPartial implementation:\nlet bn_check_mod_inv_prime_st (t: limb_t) (len: BN.meta_len t) =\n \n"
},
{
"role": "assistant",
"content": "n: lbignum t len -> a: lbignum t len\n -> Stack (limb t)\n (requires fun h -> live h n /\\ live h a /\\ disjoint n a)\n (ensures\n fun h0 r h1 -> modifies0 h0 h1 /\\ r == S.bn_check_mod_inv_prime (as_seq h0 n) (as_seq h0 a))"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: w: AlgHeap.st_wp0 a -> Type0\nEffect: Prims.Tot\n\nContext:\nmodule AlgHeap\n\n(* Essentially a copy of AlgForAll but using a heap for the state *)\n\nopen FStar.Tactics.V2\nopen FStar.List.Tot\nopen FStar.Universe\nmodule L = Lattice\nmodule Ghost = FStar.Ghost\nmodule Map = FStar.Map\nmodule T = FStar.Tactics.V2\n\ntype loc = int\ntype state = Map.t loc int\n\ntype empty : Type u#aa =\n\ntype op =\n | Read\n | Write\n | Raise\n | Other of int\n\nassume val other_inp : int -> Type u#0\nlet op_inp : op -> Type u#0 =\n function\n | Read -> unit\n | Write -> state\n | Raise -> exn\n | Other i -> other_inp i\n\nassume val other_out : int -> Type u#0\nlet op_out : op -> Type =\n function\n | Read -> state\n | Write -> unit\n | Raise -> empty\n | Other i -> other_inp i\n\nnoeq\ntype tree0 (a:Type) : Type =\n | Return : a -> tree0 a\n | Op : op:op -> i:(op_inp op) -> k:(op_out op -> tree0 a) -> tree0 a\n\ntype ops = list op\n\nlet rec abides #a (labs:ops) (f : tree0 a) : prop =\n begin match f with\n | Op a i k ->\n mem a labs /\\ (forall o. abides labs (k o))\n | Return _ -> True\n end\n\ntype tree (a:Type u#aa)\n (labs : list u#0 op)\n : Type u#aa\n =\n r:(tree0 a){abides labs r}\n\nlet rec interp_at (l1 l2 : ops) (l : op)\n : Lemma (mem l (l1@l2) == (mem l l1 || mem l l2))\n [SMTPat (mem l (l1@l2))]\n = match l1 with\n | [] -> ()\n | _::l1 -> interp_at l1 l2 l\n\nlet sublist (l1 l2 : ops) = forall x. mem x l1 ==> mem x l2\n\nlet rec sublist_at\n (l1 l2 : ops)\n : Lemma (sublist l1 (l1@l2) /\\ sublist l2 (l1@l2))\n [SMTPatOr [[SMTPat (sublist l1 (l1@l2))];\n [SMTPat (sublist l2 (l1@l2))]]]\n = match l1 with\n | [] -> ()\n | _::l1 -> sublist_at l1 l2\n\nlet sublist_at_self\n (l : ops)\n : Lemma (sublist l (l@l))\n [SMTPat (sublist l (l@l))]\n = ()\n \nlet rec abides_sublist_nopat #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c) /\\ sublist l1 l2)\n (ensures (abides l2) c)\n = match c with\n | Return _ -> ()\n | Op a i k ->\n let sub o : Lemma (abides l2 (k o)) =\n abides_sublist_nopat l1 l2 (k o)\n in\n Classical.forall_intro sub\n\nlet abides_sublist #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c) /\\ sublist l1 l2)\n (ensures (abides l2 c))\n [SMTPat (abides l2 c); SMTPat (sublist l1 l2)]\n = abides_sublist_nopat l1 l2 c\n \nlet abides_at_self #a\n (l : ops)\n (c : tree0 a)\n : Lemma (abides (l@l) c <==> abides l c)\n [SMTPat (abides (l@l) c)]\n = (* Trigger some patterns *)\n assert (sublist l (l@l));\n assert (sublist (l@l) l)\n\nlet abides_app #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c \\/ abides l2 c))\n (ensures (abides (l1@l2) c))\n [SMTPat (abides (l1@l2) c)]\n = sublist_at l1 l2\n\n(* Folding a computation tree *)\nval fold_with (#a #b:_) (#labs : ops)\n (f:tree a labs)\n (v : a -> b)\n (h: (o:op{mem o labs} -> op_inp o -> (op_out o -> b) -> b))\n : b\nlet rec fold_with #a #b #labs f v h =\n match f with\n | Return x -> v x\n | Op act i k ->\n let k' (o : op_out act) : b =\n fold_with #_ #_ #labs (k o) v h\n in\n h act i k'\n\nlet handler_ty_l (o:op) (b:Type) (labs:ops) =\n op_inp o -> (op_out o -> tree b labs) -> tree b labs\n\nlet handler_ty (labs0 : ops) (b:Type) (labs1 : ops) : Type =\n o:op{mem o labs0} -> handler_ty_l o b labs1\n\n(* The most generic handling construct, we use it to implement bind.\nIt is actually just a special case of folding. *)\nval handle_with (#a #b:_) (#labs0 #labs1 : ops)\n (f:tree a labs0)\n (v : a -> tree b labs1)\n (h: handler_ty labs0 b labs1)\n : tree b labs1\nlet handle_with f v h = fold_with f v h\n\nlet return (a:Type) (x:a)\n : tree a []\n = Return x\n\nlet bind (a b : Type)\n (#labs1 #labs2 : ops)\n (c : tree a labs1)\n (f : (x:a -> tree b labs2))\n : Tot (tree b (labs1@labs2))\n = handle_with #_ #_ #labs1 #(labs1@labs2) c f (fun act i k -> Op act i k)\n \nlet subcomp (a:Type)\n (labs1 labs2 : ops)\n (f : tree a labs1)\n : Pure (tree a labs2)\n (requires (sublist labs1 labs2))\n (ensures (fun _ -> True))\n = f\n\nlet if_then_else\n (a : Type)\n (labs1 labs2 : ops)\n (f : tree a labs1)\n (g : tree a labs2)\n (p : bool)\n : Type\n = tree a (labs1@labs2)\n\nlet _get : tree state [Read] = Op Read () Return\n\nlet _put (s:state) : tree unit [Write] = Op Write s Return\n\ntotal // need this for catch!!\nreifiable\nreflectable\neffect {\n Alg (a:Type) (_:ops)\n with {repr = tree; return; bind; subcomp; if_then_else}\n}\n\nlet get () : Alg state [Read] =\n Alg?.reflect _get\n\nlet lift_pure_eff\n (a:Type)\n (wp : pure_wp a)\n (f : unit -> PURE a wp)\n : Pure (tree a [])\n (requires (wp (fun _ -> True)))\n (ensures (fun _ -> True))\n = FStar.Monotonic.Pure.elim_pure_wp_monotonicity wp;\n Return (f ())\n\nsub_effect PURE ~> Alg = lift_pure_eff\n\nlet put (s:state) : Alg unit [Write] =\n Alg?.reflect (_put s)\n\nlet raise #a (e:exn) : Alg a [Raise] = \n Alg?.reflect (Op Raise e (fun e -> match e with))\n // funnily enough, the version below also succeeds from concluding\n // a==empty under the lambda since the context becomes inconsistent\n //Alg?.reflect (Op Raise e Return\n\ntype rwtree a = tree a [Read;Write]\n\nlet tbind : #a:_ -> #b:_ -> rwtree a -> (a -> rwtree b) -> rwtree b = fun c f -> bind _ _ c f\n\nlet st_wp0 (a:Type) : Type = state -> (a & state -> Type0) -> Type0\n\nPartial implementation:\nlet st_monotonic #a (w: st_wp0 a) : Type0 =\n \n"
},
{
"role": "assistant",
"content": "forall s0 p1 p2. (forall x s1. p1 (x, s1) ==> p2 (x, s1)) ==> w s0 p1 ==> w s0 p2"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: labs: AlgHeap.ops -> f: AlgHeap.tree0 a -> Prims.prop\nEffect: Prims.Tot\n\nContext:\nmodule AlgHeap\n\n(* Essentially a copy of AlgForAll but using a heap for the state *)\n\nopen FStar.Tactics.V2\nopen FStar.List.Tot\nopen FStar.Universe\nmodule L = Lattice\nmodule Ghost = FStar.Ghost\nmodule Map = FStar.Map\nmodule T = FStar.Tactics.V2\n\ntype loc = int\ntype state = Map.t loc int\n\ntype empty : Type u#aa =\n\ntype op =\n | Read\n | Write\n | Raise\n | Other of int\n\nassume val other_inp : int -> Type u#0\nlet op_inp : op -> Type u#0 =\n function\n | Read -> unit\n | Write -> state\n | Raise -> exn\n | Other i -> other_inp i\n\nassume val other_out : int -> Type u#0\nlet op_out : op -> Type =\n function\n | Read -> state\n | Write -> unit\n | Raise -> empty\n | Other i -> other_inp i\n\nnoeq\ntype tree0 (a:Type) : Type =\n | Return : a -> tree0 a\n | Op : op:op -> i:(op_inp op) -> k:(op_out op -> tree0 a) -> tree0 a\n\ntype ops = list op\n\nPartial implementation:\nlet rec abides #a (labs: ops) (f: tree0 a) : prop =\n \n"
},
{
"role": "assistant",
"content": "match f with\n| Op a i k -> mem a labs /\\ (forall o. abides labs (k o))\n| Return _ -> True"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: t: Hacl.Bignum.Definitions.limb_t -> len: Hacl.Bignum.meta_len t -> Type0\nEffect: Prims.Tot\n\nContext:\nmodule Hacl.Bignum.ModInv\n\nopen FStar.HyperStack\nopen FStar.HyperStack.ST\nopen FStar.Mul\n\nopen Lib.IntTypes\nopen Lib.Buffer\n\nopen Hacl.Bignum.Definitions\n\nmodule ST = FStar.HyperStack.ST\nmodule S = Hacl.Spec.Bignum.ModInv\n\nmodule BN = Hacl.Bignum\nmodule BE = Hacl.Bignum.Exponentiation\nmodule BM = Hacl.Bignum.Montgomery\nmodule SD = Hacl.Spec.Bignum.Definitions\n\n#reset-options \"--z3rlimit 50 --fuel 0 --ifuel 0\"\n\ninline_for_extraction noextract\nlet bn_check_mod_inv_prime_st (t:limb_t) (len:BN.meta_len t) =\n n:lbignum t len\n -> a:lbignum t len ->\n Stack (limb t)\n (requires fun h -> live h n /\\ live h a /\\ disjoint n a)\n (ensures fun h0 r h1 -> modifies0 h0 h1 /\\\n r == S.bn_check_mod_inv_prime (as_seq h0 n) (as_seq h0 a))\n\n\ninline_for_extraction noextract\nval bn_check_mod_inv_prime: #t:limb_t -> len:BN.meta_len t -> bn_check_mod_inv_prime_st t len\nlet bn_check_mod_inv_prime #t len n a =\n let m0 = BM.bn_check_modulus n in\n let m1 = BN.bn_is_zero_mask len a in\n let m2 = BN.bn_lt_mask len a n in\n m0 &. (lognot m1) &. m2\n\n\ninline_for_extraction noextract\nval bn_mod_inv_prime_n2:\n #t:limb_t\n -> len:BN.meta_len t\n -> n:lbignum t len\n -> res:lbignum t len ->\n Stack unit\n (requires fun h -> 1 < bn_v h n /\\\n live h n /\\ live h res /\\ disjoint res n)\n (ensures fun h0 _ h1 -> modifies (loc res) h0 h1 /\\\n as_seq h1 res == S.bn_mod_inv_prime_n2 (as_seq h0 n))\n\nlet bn_mod_inv_prime_n2 #t len n res =\n let c = BN.bn_sub1 len n (uint #t #SEC 2) res in\n LowStar.Ignore.ignore c;\n ()\n\n\ninline_for_extraction noextract\nlet bn_mod_inv_prime_precomp_st (t:limb_t) (len:BN.meta_len t) =\n n:lbignum t len\n -> mu:limb t\n -> r2:lbignum t len\n -> a:lbignum t len\n -> res:lbignum t len ->\n Stack unit\n (requires fun h ->\n live h n /\\ live h r2 /\\ live h a /\\ live h res /\\\n disjoint res n /\\ disjoint res a /\\ disjoint n a /\\\n disjoint res r2 /\\ disjoint a r2 /\\ disjoint n r2 /\\\n\n S.bn_mod_inv_prime_pre (as_seq h n) (as_seq h a) /\\\n bn_v h r2 == pow2 (2 * bits t * v len) % bn_v h n /\\\n (1 + bn_v h n * v mu) % pow2 (bits t) == 0)\n (ensures fun h0 r h1 -> modifies (loc res) h0 h1 /\\\n bn_v h1 res * bn_v h0 a % bn_v h0 n = 1)\n\n\ninline_for_extraction noextract\nval mk_bn_mod_inv_prime_precomp:\n #t:limb_t\n -> len:BN.meta_len t\n -> bn_mod_exp_precomp:BE.bn_mod_exp_precomp_st t len ->\n bn_mod_inv_prime_precomp_st t len\n\nlet mk_bn_mod_inv_prime_precomp #t len bn_mod_exp_precomp n mu r2 a res =\n let h0 = ST.get () in\n push_frame ();\n let n2 = create len (uint #t #SEC 0) in\n bn_mod_inv_prime_n2 #t len n n2;\n SD.bn_eval_bound (as_seq h0 n) (v len);\n\n bn_mod_exp_precomp n mu r2 a (size (bits t) *! len) n2 res;\n let h1 = ST.get () in\n assert (bn_v h1 res == Lib.NatMod.pow_mod #(bn_v h0 n) (bn_v h0 a) (bn_v h1 n2));\n S.mod_inv_prime_lemma (bn_v h0 n) (bn_v h0 a);\n pop_frame ()\n\nPartial implementation:\nlet bn_mod_inv_prime_st (t: limb_t) (len: BN.meta_len t) =\n \n"
},
{
"role": "assistant",
"content": "nBits: size_t -> n: lbignum t len -> a: lbignum t len -> res: lbignum t len\n -> Stack unit\n (requires\n fun h ->\n live h n /\\ live h a /\\ live h res /\\ disjoint res n /\\ disjoint res a /\\ disjoint n a /\\\n v nBits / bits t < v len /\\ pow2 (v nBits) < bn_v h n /\\\n S.bn_mod_inv_prime_pre (as_seq h n) (as_seq h a))\n (ensures fun h0 r h1 -> modifies (loc res) h0 h1 /\\ bn_v h1 res * bn_v h0 a % bn_v h0 n = 1)"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: len: Hacl.Bignum.meta_len t -> Hacl.Bignum.ModInv.bn_check_mod_inv_prime_st t len\nEffect: Prims.Tot\n\nContext:\nmodule Hacl.Bignum.ModInv\n\nopen FStar.HyperStack\nopen FStar.HyperStack.ST\nopen FStar.Mul\n\nopen Lib.IntTypes\nopen Lib.Buffer\n\nopen Hacl.Bignum.Definitions\n\nmodule ST = FStar.HyperStack.ST\nmodule S = Hacl.Spec.Bignum.ModInv\n\nmodule BN = Hacl.Bignum\nmodule BE = Hacl.Bignum.Exponentiation\nmodule BM = Hacl.Bignum.Montgomery\nmodule SD = Hacl.Spec.Bignum.Definitions\n\n#reset-options \"--z3rlimit 50 --fuel 0 --ifuel 0\"\n\ninline_for_extraction noextract\nlet bn_check_mod_inv_prime_st (t:limb_t) (len:BN.meta_len t) =\n n:lbignum t len\n -> a:lbignum t len ->\n Stack (limb t)\n (requires fun h -> live h n /\\ live h a /\\ disjoint n a)\n (ensures fun h0 r h1 -> modifies0 h0 h1 /\\\n r == S.bn_check_mod_inv_prime (as_seq h0 n) (as_seq h0 a))\n\n\ninline_for_extraction noextract\n\nPartial implementation:\nlet bn_check_mod_inv_prime #t len n a =\n \n"
},
{
"role": "assistant",
"content": "let m0 = BM.bn_check_modulus n in\nlet m1 = BN.bn_is_zero_mask len a in\nlet m2 = BN.bn_lt_mask len a n in\nm0 &. (lognot m1) &. m2"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: t: AlgHeap.rwtree a -> _: AlgHeap.state -> a * AlgHeap.state\nEffect: Prims.Tot\n\nContext:\nmodule AlgHeap\n\n(* Essentially a copy of AlgForAll but using a heap for the state *)\n\nopen FStar.Tactics.V2\nopen FStar.List.Tot\nopen FStar.Universe\nmodule L = Lattice\nmodule Ghost = FStar.Ghost\nmodule Map = FStar.Map\nmodule T = FStar.Tactics.V2\n\ntype loc = int\ntype state = Map.t loc int\n\ntype empty : Type u#aa =\n\ntype op =\n | Read\n | Write\n | Raise\n | Other of int\n\nassume val other_inp : int -> Type u#0\nlet op_inp : op -> Type u#0 =\n function\n | Read -> unit\n | Write -> state\n | Raise -> exn\n | Other i -> other_inp i\n\nassume val other_out : int -> Type u#0\nlet op_out : op -> Type =\n function\n | Read -> state\n | Write -> unit\n | Raise -> empty\n | Other i -> other_inp i\n\nnoeq\ntype tree0 (a:Type) : Type =\n | Return : a -> tree0 a\n | Op : op:op -> i:(op_inp op) -> k:(op_out op -> tree0 a) -> tree0 a\n\ntype ops = list op\n\nlet rec abides #a (labs:ops) (f : tree0 a) : prop =\n begin match f with\n | Op a i k ->\n mem a labs /\\ (forall o. abides labs (k o))\n | Return _ -> True\n end\n\ntype tree (a:Type u#aa)\n (labs : list u#0 op)\n : Type u#aa\n =\n r:(tree0 a){abides labs r}\n\nlet rec interp_at (l1 l2 : ops) (l : op)\n : Lemma (mem l (l1@l2) == (mem l l1 || mem l l2))\n [SMTPat (mem l (l1@l2))]\n = match l1 with\n | [] -> ()\n | _::l1 -> interp_at l1 l2 l\n\nlet sublist (l1 l2 : ops) = forall x. mem x l1 ==> mem x l2\n\nlet rec sublist_at\n (l1 l2 : ops)\n : Lemma (sublist l1 (l1@l2) /\\ sublist l2 (l1@l2))\n [SMTPatOr [[SMTPat (sublist l1 (l1@l2))];\n [SMTPat (sublist l2 (l1@l2))]]]\n = match l1 with\n | [] -> ()\n | _::l1 -> sublist_at l1 l2\n\nlet sublist_at_self\n (l : ops)\n : Lemma (sublist l (l@l))\n [SMTPat (sublist l (l@l))]\n = ()\n \nlet rec abides_sublist_nopat #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c) /\\ sublist l1 l2)\n (ensures (abides l2) c)\n = match c with\n | Return _ -> ()\n | Op a i k ->\n let sub o : Lemma (abides l2 (k o)) =\n abides_sublist_nopat l1 l2 (k o)\n in\n Classical.forall_intro sub\n\nlet abides_sublist #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c) /\\ sublist l1 l2)\n (ensures (abides l2 c))\n [SMTPat (abides l2 c); SMTPat (sublist l1 l2)]\n = abides_sublist_nopat l1 l2 c\n \nlet abides_at_self #a\n (l : ops)\n (c : tree0 a)\n : Lemma (abides (l@l) c <==> abides l c)\n [SMTPat (abides (l@l) c)]\n = (* Trigger some patterns *)\n assert (sublist l (l@l));\n assert (sublist (l@l) l)\n\nlet abides_app #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c \\/ abides l2 c))\n (ensures (abides (l1@l2) c))\n [SMTPat (abides (l1@l2) c)]\n = sublist_at l1 l2\n\n(* Folding a computation tree *)\nval fold_with (#a #b:_) (#labs : ops)\n (f:tree a labs)\n (v : a -> b)\n (h: (o:op{mem o labs} -> op_inp o -> (op_out o -> b) -> b))\n : b\nlet rec fold_with #a #b #labs f v h =\n match f with\n | Return x -> v x\n | Op act i k ->\n let k' (o : op_out act) : b =\n fold_with #_ #_ #labs (k o) v h\n in\n h act i k'\n\nlet handler_ty_l (o:op) (b:Type) (labs:ops) =\n op_inp o -> (op_out o -> tree b labs) -> tree b labs\n\nlet handler_ty (labs0 : ops) (b:Type) (labs1 : ops) : Type =\n o:op{mem o labs0} -> handler_ty_l o b labs1\n\n(* The most generic handling construct, we use it to implement bind.\nIt is actually just a special case of folding. *)\nval handle_with (#a #b:_) (#labs0 #labs1 : ops)\n (f:tree a labs0)\n (v : a -> tree b labs1)\n (h: handler_ty labs0 b labs1)\n : tree b labs1\nlet handle_with f v h = fold_with f v h\n\nlet return (a:Type) (x:a)\n : tree a []\n = Return x\n\nlet bind (a b : Type)\n (#labs1 #labs2 : ops)\n (c : tree a labs1)\n (f : (x:a -> tree b labs2))\n : Tot (tree b (labs1@labs2))\n = handle_with #_ #_ #labs1 #(labs1@labs2) c f (fun act i k -> Op act i k)\n \nlet subcomp (a:Type)\n (labs1 labs2 : ops)\n (f : tree a labs1)\n : Pure (tree a labs2)\n (requires (sublist labs1 labs2))\n (ensures (fun _ -> True))\n = f\n\nlet if_then_else\n (a : Type)\n (labs1 labs2 : ops)\n (f : tree a labs1)\n (g : tree a labs2)\n (p : bool)\n : Type\n = tree a (labs1@labs2)\n\nlet _get : tree state [Read] = Op Read () Return\n\nlet _put (s:state) : tree unit [Write] = Op Write s Return\n\ntotal // need this for catch!!\nreifiable\nreflectable\neffect {\n Alg (a:Type) (_:ops)\n with {repr = tree; return; bind; subcomp; if_then_else}\n}\n\nlet get () : Alg state [Read] =\n Alg?.reflect _get\n\nlet lift_pure_eff\n (a:Type)\n (wp : pure_wp a)\n (f : unit -> PURE a wp)\n : Pure (tree a [])\n (requires (wp (fun _ -> True)))\n (ensures (fun _ -> True))\n = FStar.Monotonic.Pure.elim_pure_wp_monotonicity wp;\n Return (f ())\n\nsub_effect PURE ~> Alg = lift_pure_eff\n\nlet put (s:state) : Alg unit [Write] =\n Alg?.reflect (_put s)\n\nlet raise #a (e:exn) : Alg a [Raise] = \n Alg?.reflect (Op Raise e (fun e -> match e with))\n // funnily enough, the version below also succeeds from concluding\n // a==empty under the lambda since the context becomes inconsistent\n //Alg?.reflect (Op Raise e Return\n\ntype rwtree a = tree a [Read;Write]\n\nlet tbind : #a:_ -> #b:_ -> rwtree a -> (a -> rwtree b) -> rwtree b = fun c f -> bind _ _ c f\n\nlet st_wp0 (a:Type) : Type = state -> (a & state -> Type0) -> Type0\n\nlet st_monotonic #a (w : st_wp0 a) : Type0 =\n //forall s0 p1 p2. (forall r. p1 r ==> p2 r) ==> w s0 p1 ==> w s0 p2\n // ^ this version seems to be less SMT-friendly\n forall s0 p1 p2. (forall x s1. p1 (x, s1) ==> p2 (x, s1)) ==> w s0 p1 ==> w s0 p2\n\nlet st_wp (a:Type) = wp:st_wp0 a{st_monotonic wp}\n\nunfold\nlet return_wp #a x : st_wp a = fun s0 p -> p (x, s0)\n\nunfold\nlet bind_wp #a #b (w : st_wp a) (wf : a -> st_wp b)\n : st_wp b\n = fun s0 p -> w s0 (fun (y, s1) -> wf y s1 p)\n\nunfold\nlet read_wp : st_wp state = fun s0 p -> p (s0, s0)\n\nunfold\nlet write_wp : state -> st_wp unit = fun s _ p -> p ((), s)\n\n(* Also doable with handlers *)\nlet rec interp_as_wp #a (t : rwtree a) : st_wp a =\n match t with\n | Return x -> return_wp x\n | Op Read _ k ->\n bind_wp read_wp (fun s -> interp_as_wp (k s))\n | Op Write s k ->\n bind_wp (write_wp s) (fun (o:unit) -> interp_as_wp (k o))\n\nlet rec interp_rdwr_tree #a (t : tree a [Read;Write]) (s:state) : Tot (a & state) =\n match t with\n | Return x -> (x, s)\n | Op Read _ k ->\n interp_rdwr_tree (k s) s\n | Op Write s k ->\n interp_rdwr_tree (k ()) s\n\nPartial implementation:\nlet interp_as_fun #a (t: rwtree a) : (state -> a & state) =\n \n"
},
{
"role": "assistant",
"content": "interp_rdwr_tree t"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: l1: AlgHeap.ops -> l2: AlgHeap.ops -> Prims.logical\nEffect: Prims.Tot\n\nContext:\nmodule AlgHeap\n\n(* Essentially a copy of AlgForAll but using a heap for the state *)\n\nopen FStar.Tactics.V2\nopen FStar.List.Tot\nopen FStar.Universe\nmodule L = Lattice\nmodule Ghost = FStar.Ghost\nmodule Map = FStar.Map\nmodule T = FStar.Tactics.V2\n\ntype loc = int\ntype state = Map.t loc int\n\ntype empty : Type u#aa =\n\ntype op =\n | Read\n | Write\n | Raise\n | Other of int\n\nassume val other_inp : int -> Type u#0\nlet op_inp : op -> Type u#0 =\n function\n | Read -> unit\n | Write -> state\n | Raise -> exn\n | Other i -> other_inp i\n\nassume val other_out : int -> Type u#0\nlet op_out : op -> Type =\n function\n | Read -> state\n | Write -> unit\n | Raise -> empty\n | Other i -> other_inp i\n\nnoeq\ntype tree0 (a:Type) : Type =\n | Return : a -> tree0 a\n | Op : op:op -> i:(op_inp op) -> k:(op_out op -> tree0 a) -> tree0 a\n\ntype ops = list op\n\nlet rec abides #a (labs:ops) (f : tree0 a) : prop =\n begin match f with\n | Op a i k ->\n mem a labs /\\ (forall o. abides labs (k o))\n | Return _ -> True\n end\n\ntype tree (a:Type u#aa)\n (labs : list u#0 op)\n : Type u#aa\n =\n r:(tree0 a){abides labs r}\n\nlet rec interp_at (l1 l2 : ops) (l : op)\n : Lemma (mem l (l1@l2) == (mem l l1 || mem l l2))\n [SMTPat (mem l (l1@l2))]\n = match l1 with\n | [] -> ()\n | _::l1 -> interp_at l1 l2 l\n\nPartial implementation:\nlet sublist (l1 l2: ops) =\n \n"
},
{
"role": "assistant",
"content": "forall x. mem x l1 ==> mem x l2"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: a: Type -> Type\nEffect: Prims.Tot\n\nContext:\nmodule AlgHeap\n\n(* Essentially a copy of AlgForAll but using a heap for the state *)\n\nopen FStar.Tactics.V2\nopen FStar.List.Tot\nopen FStar.Universe\nmodule L = Lattice\nmodule Ghost = FStar.Ghost\nmodule Map = FStar.Map\nmodule T = FStar.Tactics.V2\n\ntype loc = int\ntype state = Map.t loc int\n\ntype empty : Type u#aa =\n\ntype op =\n | Read\n | Write\n | Raise\n | Other of int\n\nassume val other_inp : int -> Type u#0\nlet op_inp : op -> Type u#0 =\n function\n | Read -> unit\n | Write -> state\n | Raise -> exn\n | Other i -> other_inp i\n\nassume val other_out : int -> Type u#0\nlet op_out : op -> Type =\n function\n | Read -> state\n | Write -> unit\n | Raise -> empty\n | Other i -> other_inp i\n\nnoeq\ntype tree0 (a:Type) : Type =\n | Return : a -> tree0 a\n | Op : op:op -> i:(op_inp op) -> k:(op_out op -> tree0 a) -> tree0 a\n\ntype ops = list op\n\nlet rec abides #a (labs:ops) (f : tree0 a) : prop =\n begin match f with\n | Op a i k ->\n mem a labs /\\ (forall o. abides labs (k o))\n | Return _ -> True\n end\n\ntype tree (a:Type u#aa)\n (labs : list u#0 op)\n : Type u#aa\n =\n r:(tree0 a){abides labs r}\n\nlet rec interp_at (l1 l2 : ops) (l : op)\n : Lemma (mem l (l1@l2) == (mem l l1 || mem l l2))\n [SMTPat (mem l (l1@l2))]\n = match l1 with\n | [] -> ()\n | _::l1 -> interp_at l1 l2 l\n\nlet sublist (l1 l2 : ops) = forall x. mem x l1 ==> mem x l2\n\nlet rec sublist_at\n (l1 l2 : ops)\n : Lemma (sublist l1 (l1@l2) /\\ sublist l2 (l1@l2))\n [SMTPatOr [[SMTPat (sublist l1 (l1@l2))];\n [SMTPat (sublist l2 (l1@l2))]]]\n = match l1 with\n | [] -> ()\n | _::l1 -> sublist_at l1 l2\n\nlet sublist_at_self\n (l : ops)\n : Lemma (sublist l (l@l))\n [SMTPat (sublist l (l@l))]\n = ()\n \nlet rec abides_sublist_nopat #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c) /\\ sublist l1 l2)\n (ensures (abides l2) c)\n = match c with\n | Return _ -> ()\n | Op a i k ->\n let sub o : Lemma (abides l2 (k o)) =\n abides_sublist_nopat l1 l2 (k o)\n in\n Classical.forall_intro sub\n\nlet abides_sublist #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c) /\\ sublist l1 l2)\n (ensures (abides l2 c))\n [SMTPat (abides l2 c); SMTPat (sublist l1 l2)]\n = abides_sublist_nopat l1 l2 c\n \nlet abides_at_self #a\n (l : ops)\n (c : tree0 a)\n : Lemma (abides (l@l) c <==> abides l c)\n [SMTPat (abides (l@l) c)]\n = (* Trigger some patterns *)\n assert (sublist l (l@l));\n assert (sublist (l@l) l)\n\nlet abides_app #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c \\/ abides l2 c))\n (ensures (abides (l1@l2) c))\n [SMTPat (abides (l1@l2) c)]\n = sublist_at l1 l2\n\n(* Folding a computation tree *)\nval fold_with (#a #b:_) (#labs : ops)\n (f:tree a labs)\n (v : a -> b)\n (h: (o:op{mem o labs} -> op_inp o -> (op_out o -> b) -> b))\n : b\nlet rec fold_with #a #b #labs f v h =\n match f with\n | Return x -> v x\n | Op act i k ->\n let k' (o : op_out act) : b =\n fold_with #_ #_ #labs (k o) v h\n in\n h act i k'\n\nlet handler_ty_l (o:op) (b:Type) (labs:ops) =\n op_inp o -> (op_out o -> tree b labs) -> tree b labs\n\nlet handler_ty (labs0 : ops) (b:Type) (labs1 : ops) : Type =\n o:op{mem o labs0} -> handler_ty_l o b labs1\n\n(* The most generic handling construct, we use it to implement bind.\nIt is actually just a special case of folding. *)\nval handle_with (#a #b:_) (#labs0 #labs1 : ops)\n (f:tree a labs0)\n (v : a -> tree b labs1)\n (h: handler_ty labs0 b labs1)\n : tree b labs1\nlet handle_with f v h = fold_with f v h\n\nlet return (a:Type) (x:a)\n : tree a []\n = Return x\n\nlet bind (a b : Type)\n (#labs1 #labs2 : ops)\n (c : tree a labs1)\n (f : (x:a -> tree b labs2))\n : Tot (tree b (labs1@labs2))\n = handle_with #_ #_ #labs1 #(labs1@labs2) c f (fun act i k -> Op act i k)\n \nlet subcomp (a:Type)\n (labs1 labs2 : ops)\n (f : tree a labs1)\n : Pure (tree a labs2)\n (requires (sublist labs1 labs2))\n (ensures (fun _ -> True))\n = f\n\nlet if_then_else\n (a : Type)\n (labs1 labs2 : ops)\n (f : tree a labs1)\n (g : tree a labs2)\n (p : bool)\n : Type\n = tree a (labs1@labs2)\n\nlet _get : tree state [Read] = Op Read () Return\n\nlet _put (s:state) : tree unit [Write] = Op Write s Return\n\ntotal // need this for catch!!\nreifiable\nreflectable\neffect {\n Alg (a:Type) (_:ops)\n with {repr = tree; return; bind; subcomp; if_then_else}\n}\n\nlet get () : Alg state [Read] =\n Alg?.reflect _get\n\nlet lift_pure_eff\n (a:Type)\n (wp : pure_wp a)\n (f : unit -> PURE a wp)\n : Pure (tree a [])\n (requires (wp (fun _ -> True)))\n (ensures (fun _ -> True))\n = FStar.Monotonic.Pure.elim_pure_wp_monotonicity wp;\n Return (f ())\n\nsub_effect PURE ~> Alg = lift_pure_eff\n\nlet put (s:state) : Alg unit [Write] =\n Alg?.reflect (_put s)\n\nlet raise #a (e:exn) : Alg a [Raise] = \n Alg?.reflect (Op Raise e (fun e -> match e with))\n // funnily enough, the version below also succeeds from concluding\n // a==empty under the lambda since the context becomes inconsistent\n //Alg?.reflect (Op Raise e Return\n\ntype rwtree a = tree a [Read;Write]\n\nlet tbind : #a:_ -> #b:_ -> rwtree a -> (a -> rwtree b) -> rwtree b = fun c f -> bind _ _ c f\n\nlet st_wp0 (a:Type) : Type = state -> (a & state -> Type0) -> Type0\n\nlet st_monotonic #a (w : st_wp0 a) : Type0 =\n //forall s0 p1 p2. (forall r. p1 r ==> p2 r) ==> w s0 p1 ==> w s0 p2\n // ^ this version seems to be less SMT-friendly\n forall s0 p1 p2. (forall x s1. p1 (x, s1) ==> p2 (x, s1)) ==> w s0 p1 ==> w s0 p2\n\nPartial implementation:\nlet st_wp (a: Type) =\n \n"
},
{
"role": "assistant",
"content": "wp: st_wp0 a {st_monotonic wp}"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: w1: AlgHeap.st_wp a -> w2: AlgHeap.st_wp a -> Type0\nEffect: Prims.Tot\n\nContext:\nmodule AlgHeap\n\n(* Essentially a copy of AlgForAll but using a heap for the state *)\n\nopen FStar.Tactics.V2\nopen FStar.List.Tot\nopen FStar.Universe\nmodule L = Lattice\nmodule Ghost = FStar.Ghost\nmodule Map = FStar.Map\nmodule T = FStar.Tactics.V2\n\ntype loc = int\ntype state = Map.t loc int\n\ntype empty : Type u#aa =\n\ntype op =\n | Read\n | Write\n | Raise\n | Other of int\n\nassume val other_inp : int -> Type u#0\nlet op_inp : op -> Type u#0 =\n function\n | Read -> unit\n | Write -> state\n | Raise -> exn\n | Other i -> other_inp i\n\nassume val other_out : int -> Type u#0\nlet op_out : op -> Type =\n function\n | Read -> state\n | Write -> unit\n | Raise -> empty\n | Other i -> other_inp i\n\nnoeq\ntype tree0 (a:Type) : Type =\n | Return : a -> tree0 a\n | Op : op:op -> i:(op_inp op) -> k:(op_out op -> tree0 a) -> tree0 a\n\ntype ops = list op\n\nlet rec abides #a (labs:ops) (f : tree0 a) : prop =\n begin match f with\n | Op a i k ->\n mem a labs /\\ (forall o. abides labs (k o))\n | Return _ -> True\n end\n\ntype tree (a:Type u#aa)\n (labs : list u#0 op)\n : Type u#aa\n =\n r:(tree0 a){abides labs r}\n\nlet rec interp_at (l1 l2 : ops) (l : op)\n : Lemma (mem l (l1@l2) == (mem l l1 || mem l l2))\n [SMTPat (mem l (l1@l2))]\n = match l1 with\n | [] -> ()\n | _::l1 -> interp_at l1 l2 l\n\nlet sublist (l1 l2 : ops) = forall x. mem x l1 ==> mem x l2\n\nlet rec sublist_at\n (l1 l2 : ops)\n : Lemma (sublist l1 (l1@l2) /\\ sublist l2 (l1@l2))\n [SMTPatOr [[SMTPat (sublist l1 (l1@l2))];\n [SMTPat (sublist l2 (l1@l2))]]]\n = match l1 with\n | [] -> ()\n | _::l1 -> sublist_at l1 l2\n\nlet sublist_at_self\n (l : ops)\n : Lemma (sublist l (l@l))\n [SMTPat (sublist l (l@l))]\n = ()\n \nlet rec abides_sublist_nopat #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c) /\\ sublist l1 l2)\n (ensures (abides l2) c)\n = match c with\n | Return _ -> ()\n | Op a i k ->\n let sub o : Lemma (abides l2 (k o)) =\n abides_sublist_nopat l1 l2 (k o)\n in\n Classical.forall_intro sub\n\nlet abides_sublist #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c) /\\ sublist l1 l2)\n (ensures (abides l2 c))\n [SMTPat (abides l2 c); SMTPat (sublist l1 l2)]\n = abides_sublist_nopat l1 l2 c\n \nlet abides_at_self #a\n (l : ops)\n (c : tree0 a)\n : Lemma (abides (l@l) c <==> abides l c)\n [SMTPat (abides (l@l) c)]\n = (* Trigger some patterns *)\n assert (sublist l (l@l));\n assert (sublist (l@l) l)\n\nlet abides_app #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c \\/ abides l2 c))\n (ensures (abides (l1@l2) c))\n [SMTPat (abides (l1@l2) c)]\n = sublist_at l1 l2\n\n(* Folding a computation tree *)\nval fold_with (#a #b:_) (#labs : ops)\n (f:tree a labs)\n (v : a -> b)\n (h: (o:op{mem o labs} -> op_inp o -> (op_out o -> b) -> b))\n : b\nlet rec fold_with #a #b #labs f v h =\n match f with\n | Return x -> v x\n | Op act i k ->\n let k' (o : op_out act) : b =\n fold_with #_ #_ #labs (k o) v h\n in\n h act i k'\n\nlet handler_ty_l (o:op) (b:Type) (labs:ops) =\n op_inp o -> (op_out o -> tree b labs) -> tree b labs\n\nlet handler_ty (labs0 : ops) (b:Type) (labs1 : ops) : Type =\n o:op{mem o labs0} -> handler_ty_l o b labs1\n\n(* The most generic handling construct, we use it to implement bind.\nIt is actually just a special case of folding. *)\nval handle_with (#a #b:_) (#labs0 #labs1 : ops)\n (f:tree a labs0)\n (v : a -> tree b labs1)\n (h: handler_ty labs0 b labs1)\n : tree b labs1\nlet handle_with f v h = fold_with f v h\n\nlet return (a:Type) (x:a)\n : tree a []\n = Return x\n\nlet bind (a b : Type)\n (#labs1 #labs2 : ops)\n (c : tree a labs1)\n (f : (x:a -> tree b labs2))\n : Tot (tree b (labs1@labs2))\n = handle_with #_ #_ #labs1 #(labs1@labs2) c f (fun act i k -> Op act i k)\n \nlet subcomp (a:Type)\n (labs1 labs2 : ops)\n (f : tree a labs1)\n : Pure (tree a labs2)\n (requires (sublist labs1 labs2))\n (ensures (fun _ -> True))\n = f\n\nlet if_then_else\n (a : Type)\n (labs1 labs2 : ops)\n (f : tree a labs1)\n (g : tree a labs2)\n (p : bool)\n : Type\n = tree a (labs1@labs2)\n\nlet _get : tree state [Read] = Op Read () Return\n\nlet _put (s:state) : tree unit [Write] = Op Write s Return\n\ntotal // need this for catch!!\nreifiable\nreflectable\neffect {\n Alg (a:Type) (_:ops)\n with {repr = tree; return; bind; subcomp; if_then_else}\n}\n\nlet get () : Alg state [Read] =\n Alg?.reflect _get\n\nlet lift_pure_eff\n (a:Type)\n (wp : pure_wp a)\n (f : unit -> PURE a wp)\n : Pure (tree a [])\n (requires (wp (fun _ -> True)))\n (ensures (fun _ -> True))\n = FStar.Monotonic.Pure.elim_pure_wp_monotonicity wp;\n Return (f ())\n\nsub_effect PURE ~> Alg = lift_pure_eff\n\nlet put (s:state) : Alg unit [Write] =\n Alg?.reflect (_put s)\n\nlet raise #a (e:exn) : Alg a [Raise] = \n Alg?.reflect (Op Raise e (fun e -> match e with))\n // funnily enough, the version below also succeeds from concluding\n // a==empty under the lambda since the context becomes inconsistent\n //Alg?.reflect (Op Raise e Return\n\ntype rwtree a = tree a [Read;Write]\n\nlet tbind : #a:_ -> #b:_ -> rwtree a -> (a -> rwtree b) -> rwtree b = fun c f -> bind _ _ c f\n\nlet st_wp0 (a:Type) : Type = state -> (a & state -> Type0) -> Type0\n\nlet st_monotonic #a (w : st_wp0 a) : Type0 =\n //forall s0 p1 p2. (forall r. p1 r ==> p2 r) ==> w s0 p1 ==> w s0 p2\n // ^ this version seems to be less SMT-friendly\n forall s0 p1 p2. (forall x s1. p1 (x, s1) ==> p2 (x, s1)) ==> w s0 p1 ==> w s0 p2\n\nlet st_wp (a:Type) = wp:st_wp0 a{st_monotonic wp}\n\nunfold\nlet return_wp #a x : st_wp a = fun s0 p -> p (x, s0)\n\nunfold\nlet bind_wp #a #b (w : st_wp a) (wf : a -> st_wp b)\n : st_wp b\n = fun s0 p -> w s0 (fun (y, s1) -> wf y s1 p)\n\nunfold\nlet read_wp : st_wp state = fun s0 p -> p (s0, s0)\n\nunfold\nlet write_wp : state -> st_wp unit = fun s _ p -> p ((), s)\n\n(* Also doable with handlers *)\nlet rec interp_as_wp #a (t : rwtree a) : st_wp a =\n match t with\n | Return x -> return_wp x\n | Op Read _ k ->\n bind_wp read_wp (fun s -> interp_as_wp (k s))\n | Op Write s k ->\n bind_wp (write_wp s) (fun (o:unit) -> interp_as_wp (k o))\n\nlet rec interp_rdwr_tree #a (t : tree a [Read;Write]) (s:state) : Tot (a & state) =\n match t with\n | Return x -> (x, s)\n | Op Read _ k ->\n interp_rdwr_tree (k s) s\n | Op Write s k ->\n interp_rdwr_tree (k ()) s\n\nlet interp_as_fun #a (t : rwtree a) : (state -> a & state) =\n interp_rdwr_tree t\n\n(* Bug: defining this as a FStar.Preorder.preorder\ncauses stupid failures ahead *)\n\nPartial implementation:\nlet stronger w1 w2 =\n \n"
},
{
"role": "assistant",
"content": "forall p s. w1 p s ==> w2 p s"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n out: Vale.Inline.X64.Fadd_inline.u256 ->\n f1: Vale.Inline.X64.Fadd_inline.u256 ->\n f2: Vale.Inline.X64.Fadd_inline.u256\n -> FStar.HyperStack.ST.Stack Prims.unit\nEffect: FStar.HyperStack.ST.Stack\n\nContext:\nmodule Vale.Inline.X64.Fadd_inline\n\nopen FStar.Mul\nopen FStar.HyperStack.ST\nmodule HS = FStar.HyperStack\nmodule B = LowStar.Buffer\nmodule DV = LowStar.BufferView.Down\nopen Vale.Def.Types_s\n\nopen Vale.Interop.Base\nmodule IX64 = Vale.Interop.X64\nmodule VSig = Vale.AsLowStar.ValeSig\nmodule LSig = Vale.AsLowStar.LowStarSig\nmodule ME = Vale.X64.Memory\nmodule V = Vale.X64.Decls\nmodule IA = Vale.Interop.Assumptions\nmodule W = Vale.AsLowStar.Wrapper\nopen Vale.X64.MemoryAdapters\nmodule VS = Vale.X64.State\nmodule MS = Vale.X64.Machine_s\nmodule PR = Vale.X64.Print_Inline_s\n\nmodule FU = Vale.Curve25519.X64.FastUtil\nmodule FH = Vale.Curve25519.X64.FastHybrid\nmodule FW = Vale.Curve25519.X64.FastWide\n\nlet uint64 = UInt64.t\n\n(* A little utility to trigger normalization in types *)\nlet as_t (#a:Type) (x:normal a) : a = x\nlet as_normal_t (#a:Type) (x:a) : normal a = x\n\n[@__reduce__]\nlet b64 = buf_t TUInt64 TUInt64\n[@__reduce__]\nlet t64_mod = TD_Buffer TUInt64 TUInt64 default_bq\n[@__reduce__]\nlet t64_no_mod = TD_Buffer TUInt64 TUInt64 ({modified=false; strict_disjointness=false; taint=MS.Secret})\n[@__reduce__]\nlet tuint64 = TD_Base TUInt64\n\n[@__reduce__]\nlet dom: IX64.arity_ok 3 td =\n let y = [t64_mod; t64_no_mod; tuint64] in\n assert_norm (List.length y = 3);\n y\n\n(* Need to rearrange the order of arguments *)\n[@__reduce__]\nlet add1_pre : VSig.vale_pre dom =\n fun (c:V.va_code)\n (out:b64)\n (f1:b64)\n (f2:uint64)\n (va_s0:V.va_state) ->\n FU.va_req_Fast_add1 c va_s0\n (as_vale_buffer out) (as_vale_buffer f1) (UInt64.v f2)\n\n[@__reduce__]\nlet add1_post : VSig.vale_post dom =\n fun (c:V.va_code)\n (out:b64)\n (f1:b64)\n (f2:uint64)\n (va_s0:V.va_state)\n (va_s1:V.va_state)\n (f:V.va_fuel) ->\n FU.va_ens_Fast_add1 c va_s0 (as_vale_buffer out) (as_vale_buffer f1) (UInt64.v f2) va_s1 f\n\n#set-options \"--z3rlimit 50\"\n\nlet add1_regs_modified: MS.reg_64 -> bool = fun (r:MS.reg_64) ->\n let open MS in\n if r = rRax || r = rRdx || r = rR8 || r = rR9 || r = rR10 || r = rR11 then true\n else false\n\nlet add1_xmms_modified = fun _ -> false\n\n[@__reduce__]\nlet add1_lemma'\n (code:V.va_code)\n (_win:bool)\n (out:b64)\n (f1:b64)\n (f2:uint64)\n (va_s0:V.va_state)\n : Ghost (V.va_state & V.va_fuel)\n (requires\n add1_pre code out f1 f2 va_s0)\n (ensures (fun (va_s1, f) ->\n V.eval_code code va_s0 f va_s1 /\\\n VSig.vale_calling_conventions va_s0 va_s1 add1_regs_modified add1_xmms_modified /\\\n add1_post code out f1 f2 va_s0 va_s1 f /\\\n ME.buffer_readable (VS.vs_get_vale_heap va_s1) (as_vale_buffer f1) /\\\n ME.buffer_readable (VS.vs_get_vale_heap va_s1) (as_vale_buffer out) /\\\n ME.buffer_writeable (as_vale_buffer out) /\\\n ME.buffer_writeable (as_vale_buffer f1) /\\\n ME.modifies (ME.loc_union (ME.loc_buffer (as_vale_buffer out))\n ME.loc_none) (VS.vs_get_vale_heap va_s0) (VS.vs_get_vale_heap va_s1)\n )) =\n let va_s1, f = FU.va_lemma_Fast_add1 code va_s0 (as_vale_buffer out) (as_vale_buffer f1) (UInt64.v f2) in\n Vale.AsLowStar.MemoryHelpers.buffer_writeable_reveal ME.TUInt64 ME.TUInt64 out;\n Vale.AsLowStar.MemoryHelpers.buffer_writeable_reveal ME.TUInt64 ME.TUInt64 f1;\n (va_s1, f)\n\n(* Prove that add1_lemma' has the required type *)\nlet add1_lemma = as_t #(VSig.vale_sig add1_regs_modified add1_xmms_modified add1_pre add1_post) add1_lemma'\n\nlet code_add1 = FU.va_code_Fast_add1 ()\n\nlet of_reg (r:MS.reg_64) : option (IX64.reg_nat 3) = match r with\n | 5 -> Some 0 // rdi\n | 4 -> Some 1 // rsi\n | 3 -> Some 2 // rdx\n | _ -> None\n\nlet of_arg (i:IX64.reg_nat 3) : MS.reg_64 = match i with\n | 0 -> MS.rRdi\n | 1 -> MS.rRsi\n | 2 -> MS.rRdx\n\nlet arg_reg : IX64.arg_reg_relation 3 = IX64.Rel of_reg of_arg\n\n(* Here's the type expected for the add1 wrapper *)\n[@__reduce__]\nlet lowstar_add1_t =\n assert_norm (List.length dom + List.length ([]<:list arg) <= 3);\n IX64.as_lowstar_sig_t_weak\n 3\n arg_reg\n add1_regs_modified\n add1_xmms_modified\n code_add1\n dom\n []\n _\n _\n // The boolean here doesn't matter\n (W.mk_prediction code_add1 dom [] (add1_lemma code_add1 IA.win))\n\n(* And here's the add1 wrapper itself *)\nlet lowstar_add1 : lowstar_add1_t =\n assert_norm (List.length dom + List.length ([]<:list arg) <= 3);\n IX64.wrap_weak\n 3\n arg_reg\n add1_regs_modified\n add1_xmms_modified\n code_add1\n dom\n (W.mk_prediction code_add1 dom [] (add1_lemma code_add1 IA.win))\n\nlet lowstar_add1_normal_t : normal lowstar_add1_t\n = as_normal_t #lowstar_add1_t lowstar_add1\n\nopen Vale.AsLowStar.MemoryHelpers\n\nlet add_scalar out f1 f2\n = DV.length_eq (get_downview out);\n DV.length_eq (get_downview f1);\n let (x, _) = lowstar_add1_normal_t out f1 f2 () in\n x\n\nlet add1_comments : list string =\n [\"Computes the addition of four-element f1 with value in f2\"; \"and returns the carry (if any)\"]\n\nlet add1_names (n:nat) =\n match n with\n | 0 -> \"out\"\n | 1 -> \"f1\"\n | 2 -> \"f2\"\n | _ -> \"\"\n\nlet add1_code_inline () : FStar.All.ML int =\n PR.print_inline \"add_scalar\" 0 None (List.length dom) dom add1_names code_add1 of_arg add1_regs_modified add1_comments\n\n\n[@__reduce__]\nlet fadd_dom: IX64.arity_ok_stdcall td =\n let y = [t64_mod; t64_no_mod; t64_no_mod] in\n assert_norm (List.length y = 3);\n y\n\n(* Need to rearrange the order of arguments *)\n[@__reduce__]\nlet fadd_pre : VSig.vale_pre fadd_dom =\n fun (c:V.va_code)\n (out:b64)\n (f1:b64)\n (f2:b64)\n (va_s0:V.va_state) ->\n FH.va_req_Fadd c va_s0\n (as_vale_buffer out) (as_vale_buffer f1) (as_vale_buffer f2)\n\n[@__reduce__]\nlet fadd_post : VSig.vale_post fadd_dom =\n fun (c:V.va_code)\n (out:b64)\n (f1:b64)\n (f2:b64)\n (va_s0:V.va_state)\n (va_s1:V.va_state)\n (f:V.va_fuel) ->\n FH.va_ens_Fadd c va_s0 (as_vale_buffer out) (as_vale_buffer f1) (as_vale_buffer f2) va_s1 f\n\n#set-options \"--z3rlimit 50\"\n\nlet fadd_regs_modified: MS.reg_64 -> bool = fun (r:MS.reg_64) ->\n let open MS in\n if r = rRax || r = rRcx || r = rRdx || r = rR8 || r = rR9 || r = rR10 || r = rR11 then true\n else false\n\nlet fadd_xmms_modified = fun _ -> false\n\n[@__reduce__]\nlet fadd_lemma'\n (code:V.va_code)\n (_win:bool)\n (out:b64)\n (f1:b64)\n (f2:b64)\n (va_s0:V.va_state)\n : Ghost (V.va_state & V.va_fuel)\n (requires\n fadd_pre code out f1 f2 va_s0)\n (ensures (fun (va_s1, f) ->\n V.eval_code code va_s0 f va_s1 /\\\n VSig.vale_calling_conventions va_s0 va_s1 fadd_regs_modified fadd_xmms_modified /\\\n fadd_post code out f1 f2 va_s0 va_s1 f /\\\n ME.buffer_readable (VS.vs_get_vale_heap va_s1) (as_vale_buffer out) /\\\n ME.buffer_readable (VS.vs_get_vale_heap va_s1) (as_vale_buffer f1) /\\\n ME.buffer_readable (VS.vs_get_vale_heap va_s1) (as_vale_buffer f2) /\\\n ME.buffer_writeable (as_vale_buffer out) /\\\n ME.buffer_writeable (as_vale_buffer f1) /\\\n ME.buffer_writeable (as_vale_buffer f2) /\\\n ME.modifies (ME.loc_union (ME.loc_buffer (as_vale_buffer out))\n ME.loc_none) (VS.vs_get_vale_heap va_s0) (VS.vs_get_vale_heap va_s1)\n )) =\n let va_s1, f = FH.va_lemma_Fadd code va_s0 (as_vale_buffer out) (as_vale_buffer f1) (as_vale_buffer f2) in\n Vale.AsLowStar.MemoryHelpers.buffer_writeable_reveal ME.TUInt64 ME.TUInt64 out;\n Vale.AsLowStar.MemoryHelpers.buffer_writeable_reveal ME.TUInt64 ME.TUInt64 f1;\n Vale.AsLowStar.MemoryHelpers.buffer_writeable_reveal ME.TUInt64 ME.TUInt64 f2;\n (va_s1, f)\n\n(* Prove that add1_lemma' has the required type *)\nlet fadd_lemma = as_t #(VSig.vale_sig fadd_regs_modified fadd_xmms_modified fadd_pre fadd_post) fadd_lemma'\n\nlet code_Fadd = FH.va_code_Fadd ()\n\n(* Here's the type expected for the fadd wrapper *)\n[@__reduce__]\nlet lowstar_fadd_t =\n assert_norm (List.length fadd_dom + List.length ([]<:list arg) <= 3);\n IX64.as_lowstar_sig_t_weak\n 3\n arg_reg\n fadd_regs_modified\n fadd_xmms_modified\n code_Fadd\n fadd_dom\n []\n _\n _\n // The boolean here doesn't matter\n (W.mk_prediction code_Fadd fadd_dom [] (fadd_lemma code_Fadd IA.win))\n\n(* And here's the fadd wrapper itself *)\nlet lowstar_fadd : lowstar_fadd_t =\n assert_norm (List.length fadd_dom + List.length ([]<:list arg) <= 3);\n IX64.wrap_weak\n 3\n arg_reg\n fadd_regs_modified\n fadd_xmms_modified\n code_Fadd\n fadd_dom\n (W.mk_prediction code_Fadd fadd_dom [] (fadd_lemma code_Fadd IA.win))\n\nlet lowstar_fadd_normal_t : normal lowstar_fadd_t\n = as_normal_t #lowstar_fadd_t lowstar_fadd\n\nPartial implementation:\nlet fadd out f1 f2 =\n \n"
},
{
"role": "assistant",
"content": "DV.length_eq (get_downview out);\nDV.length_eq (get_downview f1);\nDV.length_eq (get_downview f2);\nlet x, _ = lowstar_fadd_normal_t out f1 f2 () in\n()"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n alg: Vale.AES.AES_common_s.algorithm ->\n offset: Prims.int ->\n auth_b: Vale.X64.Memory.buffer128 ->\n abytes_b: Vale.X64.Memory.buffer128 ->\n in128x6_b: Vale.X64.Memory.buffer128 ->\n out128x6_b: Vale.X64.Memory.buffer128 ->\n in128_b: Vale.X64.Memory.buffer128 ->\n out128_b: Vale.X64.Memory.buffer128 ->\n inout_b: Vale.X64.Memory.buffer128 ->\n iv_b: Vale.X64.Memory.buffer128 ->\n iv: Vale.AES.GCM_s.supported_iv_LE ->\n scratch_b: Vale.X64.Memory.buffer128 ->\n key: FStar.Seq.Base.seq Vale.X64.Memory.nat32 ->\n round_keys: FStar.Seq.Base.seq Vale.X64.Decls.quad32 ->\n keys_b: Vale.X64.Memory.buffer128 ->\n hkeys_b: Vale.X64.Memory.buffer128 ->\n expected_tag: FStar.Seq.Base.seq Vale.X64.Memory.nat8 ->\n va_s0: Vale.X64.Decls.va_state ->\n va_k: (_: Vale.X64.Decls.va_state -> _: Prims.unit -> Type0)\n -> Type0\nEffect: Prims.Tot\n\nContext:\nmodule Vale.AES.X64.GCMdecryptOpt\nopen Vale.Def.Prop_s\nopen Vale.Def.Opaque_s\nopen FStar.Seq\nopen Vale.Def.Words_s\nopen Vale.Def.Words.Seq_s\nopen Vale.Def.Types_s\nopen Vale.Arch.Types\nopen Vale.Arch.HeapImpl\nopen Vale.AES.AES_s\nopen Vale.AES.GCTR_s\nopen Vale.AES.GCTR\nopen Vale.AES.GCM\nopen Vale.AES.GHash_s\nopen Vale.AES.GHash\nopen Vale.AES.GCM_s\nopen Vale.AES.X64.AES\nopen Vale.AES.GF128_s\nopen Vale.AES.GF128\nopen Vale.Poly1305.Math\nopen Vale.AES.GCM_helpers\nopen Vale.AES.X64.GHash\nopen Vale.AES.X64.GCTR\nopen Vale.X64.Machine_s\nopen Vale.X64.Memory\nopen Vale.X64.Stack_i\nopen Vale.X64.State\nopen Vale.X64.Decls\nopen Vale.X64.InsBasic\nopen Vale.X64.InsMem\nopen Vale.X64.InsVector\nopen Vale.X64.InsStack\nopen Vale.X64.InsAes\nopen Vale.X64.QuickCode\nopen Vale.X64.QuickCodes\nopen Vale.AES.X64.GF128_Mul\nopen Vale.X64.Stack\nopen Vale.X64.CPU_Features_s\nopen Vale.Math.Poly2.Bits_s\nopen Vale.AES.X64.AESopt\nopen Vale.AES.X64.AESGCM\nopen Vale.AES.X64.AESopt2\nopen Vale.Lib.Meta\nopen Vale.AES.X64.GCMencryptOpt\nopen Vale.AES.OptPublic\nopen Vale.Lib.Basic\n#reset-options \"--z3rlimit 20 --max_ifuel 0\"\n//-- Gcm_extra_bytes\n\nval va_code_Gcm_extra_bytes : alg:algorithm -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_extra_bytes alg =\n (va_Block (va_CCons (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0)\n (va_op_reg_opr64_reg64 rRax) 0 Secret) (va_CCons (va_code_Mov128 (va_op_xmm_xmm 10)\n (va_op_xmm_xmm 0)) (va_CCons (va_code_Ghash_extra_bytes ()) (va_CCons (va_code_Mov128\n (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_CCons (va_code_Pshufb (va_op_xmm_xmm 0)\n (va_op_xmm_xmm 9)) (va_CCons (va_code_AESEncryptBlock alg) (va_CCons (va_code_Pxor\n (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (va_CCons (va_code_Store128_buffer\n (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax) (va_op_xmm_xmm 10) 0 Secret)\n (va_CNil ()))))))))))\n\nval va_codegen_success_Gcm_extra_bytes : alg:algorithm -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_extra_bytes alg =\n (va_pbool_and (va_codegen_success_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0)\n (va_op_reg_opr64_reg64 rRax) 0 Secret) (va_pbool_and (va_codegen_success_Mov128 (va_op_xmm_xmm\n 10) (va_op_xmm_xmm 0)) (va_pbool_and (va_codegen_success_Ghash_extra_bytes ()) (va_pbool_and\n (va_codegen_success_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_pbool_and\n (va_codegen_success_Pshufb (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (va_pbool_and\n (va_codegen_success_AESEncryptBlock alg) (va_pbool_and (va_codegen_success_Pxor (va_op_xmm_xmm\n 10) (va_op_xmm_xmm 0)) (va_pbool_and (va_codegen_success_Store128_buffer\n (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax) (va_op_xmm_xmm 10) 0 Secret)\n (va_ttrue ())))))))))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_extra_bytes (va_mods:va_mods_t) (alg:algorithm) (inout_b:buffer128) (key:(seq\n nat32)) (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (total_bytes:nat)\n (old_hash:quad32) (completed_quads:(seq quad32)) (h_LE:quad32) : (va_quickCode unit\n (va_code_Gcm_extra_bytes alg)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in let (len:(va_int_range\n 1 1)) = 1 in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 188 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64\n rRax) 0 Secret inout_b 0) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 189 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov128 (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (fun (va_s:va_state) _ -> let\n (hash_input:quad32) = va_get_xmm 0 va_s in va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 193 column 22 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Ghash_extra_bytes hkeys_b total_bytes old_hash h_LE completed_quads) (fun\n (va_s:va_state) _ -> va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 194 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (FStar.Seq.Base.equal #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s)\n inout_b) (FStar.Seq.Base.create #quad32 1 hash_input)) (let (snap:(FStar.Seq.Base.seq\n Vale.X64.Decls.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s) inout_b in va_QSeq\n va_range1\n \"***** PRECONDITION NOT MET AT line 198 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 11)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 199 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Pshufb (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (fun (va_s:va_state) _ -> va_QBind\n va_range1\n \"***** PRECONDITION NOT MET AT line 200 column 20 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_AESEncryptBlock alg (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 11 va_s)) key\n round_keys keys_b) (fun (va_s:va_state) _ -> va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 201 column 26 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.AES_s.aes_encrypt_LE_reveal ()) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 204 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Pxor (va_op_xmm_xmm 10) (va_op_xmm_xmm 0)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 205 column 20 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Store128_buffer (va_op_heaplet_mem_heaplet 5) (va_op_reg_opr64_reg64 rRax)\n (va_op_xmm_xmm 10) 0 Secret inout_b 0) (fun (va_s:va_state) _ -> va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 207 column 24 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.gctr_partial_reveal ()) (va_QEmpty (()))))))))))))))\n\n\nval va_lemma_Gcm_extra_bytes : va_b0:va_code -> va_s0:va_state -> alg:algorithm ->\n inout_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128 -> total_bytes:nat -> old_hash:quad32 -> completed_quads:(seq quad32) ->\n h_LE:quad32\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_extra_bytes alg) va_s0 /\\ va_get_ok va_s0 /\\ (let\n (len:(va_int_range 1 1)) = 1 in sse_enabled /\\ Vale.X64.Decls.buffers_disjoint128 keys_b\n inout_b /\\ Vale.X64.Decls.buffers_disjoint128 hkeys_b inout_b /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) (va_get_reg64 rRax va_s0) inout_b\n len (va_get_mem_layout va_s0) Secret /\\ len == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 inout_b /\\ va_get_xmm 9 va_s0 == Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\ aes_reqs alg key round_keys\n keys_b (va_get_reg64 rR8 va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\\n pclmulqdq_enabled /\\ Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0\n va_s0) hkeys_b) (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret /\\ va_get_xmm 8 va_s0 == Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.AES.GHash.ghash_incremental0 h_LE old_hash completed_quads) /\\ FStar.Seq.Base.length\n #quad32 completed_quads == total_bytes `op_Division` 16 /\\ total_bytes < 16 `op_Multiply`\n FStar.Seq.Base.length #quad32 completed_quads + 16 /\\ va_get_reg64 rR10 va_s0 == total_bytes\n `op_Modulus` 16 /\\ total_bytes `op_Modulus` 16 =!= 0 /\\ (0 < total_bytes /\\ total_bytes < 16\n `op_Multiply` Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes) /\\ 16 `op_Multiply`\n (Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes - 1) < total_bytes)))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (let (len:(va_int_range 1 1)) = 1 in Vale.X64.Decls.modifies_buffer128 inout_b\n (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5 va_sM) /\\ Vale.AES.GCTR.gctr_partial alg len\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b) key (va_get_xmm 11 va_s0) /\\ (let raw_quads =\n FStar.Seq.Base.append #quad32 completed_quads (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0)\n inout_b) in let input_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads) 0 total_bytes in let padded_bytes =\n Vale.AES.GCTR_s.pad_to_128_bits input_bytes in let input_quads =\n Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in l_and (FStar.Seq.Base.length\n #Vale.Def.Types_s.quad32 input_quads > 0) (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8\n va_sM) == Vale.AES.GHash.ghash_incremental h_LE old_hash input_quads))) /\\ va_state_eq va_sM\n (va_update_flags va_sM (va_update_mem_heaplet 5 va_sM (va_update_xmm 10 va_sM (va_update_xmm 8\n va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5 va_sM (va_update_xmm 4\n va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1 va_sM (va_update_xmm 0\n va_sM (va_update_reg64 rR11 va_sM (va_update_reg64 rRcx va_sM (va_update_ok va_sM\n (va_update_mem va_sM va_s0))))))))))))))))))\n[@\"opaque_to_smt\"]\nlet va_lemma_Gcm_extra_bytes va_b0 va_s0 alg inout_b key round_keys keys_b hkeys_b total_bytes\n old_hash completed_quads h_LE =\n let (va_mods:va_mods_t) = [va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8;\n va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm\n 1; va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_ok; va_Mod_mem] in\n let va_qc = va_qcode_Gcm_extra_bytes va_mods alg inout_b key round_keys keys_b hkeys_b\n total_bytes old_hash completed_quads h_LE in\n let (va_sM, va_fM, va_g) = va_wp_sound_code_norm (va_code_Gcm_extra_bytes alg) va_qc va_s0 (fun\n va_s0 va_sM va_g -> let () = va_g in label va_range1\n \"***** POSTCONDITION NOT MET AT line 121 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\ (let (len:(va_int_range 1 1)) = 1 in label va_range1\n \"***** POSTCONDITION NOT MET AT line 174 column 55 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 177 column 95 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.AES.GCTR.gctr_partial alg len (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM) inout_b) key (va_get_xmm 11 va_s0)) /\\ label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 180 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let raw_quads = FStar.Seq.Base.append #quad32 completed_quads (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 181 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let input_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads) 0 total_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 182 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let padded_bytes = Vale.AES.GCTR_s.pad_to_128_bits input_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 183 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let input_quads = Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 186 column 59 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (l_and (FStar.Seq.Base.length #Vale.Def.Types_s.quad32 input_quads > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE old_hash input_quads)))))))) in\n assert_norm (va_qc.mods == va_mods);\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm\n 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1;\n va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_ok; va_Mod_mem]) va_sM va_s0;\n (va_sM, va_fM)\n\n[@ va_qattr]\nlet va_wp_Gcm_extra_bytes (alg:algorithm) (inout_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (total_bytes:nat) (old_hash:quad32)\n (completed_quads:(seq quad32)) (h_LE:quad32) (va_s0:va_state) (va_k:(va_state -> unit -> Type0))\n : Type0 =\n (va_get_ok va_s0 /\\ (let (len:(va_int_range 1 1)) = 1 in sse_enabled /\\\n Vale.X64.Decls.buffers_disjoint128 keys_b inout_b /\\ Vale.X64.Decls.buffers_disjoint128 hkeys_b\n inout_b /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) (va_get_reg64 rRax\n va_s0) inout_b len (va_get_mem_layout va_s0) Secret /\\ len == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 inout_b /\\ va_get_xmm 9 va_s0 == Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\ aes_reqs alg key round_keys\n keys_b (va_get_reg64 rR8 va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\\n pclmulqdq_enabled /\\ Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0\n va_s0) hkeys_b) (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret /\\ va_get_xmm 8 va_s0 == Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.AES.GHash.ghash_incremental0 h_LE old_hash completed_quads) /\\ FStar.Seq.Base.length\n #quad32 completed_quads == total_bytes `op_Division` 16 /\\ total_bytes < 16 `op_Multiply`\n FStar.Seq.Base.length #quad32 completed_quads + 16 /\\ va_get_reg64 rR10 va_s0 == total_bytes\n `op_Modulus` 16 /\\ total_bytes `op_Modulus` 16 =!= 0 /\\ (0 < total_bytes /\\ total_bytes < 16\n `op_Multiply` Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes) /\\ 16 `op_Multiply`\n (Vale.AES.GCM_helpers.bytes_to_quad_size total_bytes - 1) < total_bytes) /\\ (forall\n (va_x_mem:vale_heap) (va_x_rcx:nat64) (va_x_r11:nat64) (va_x_xmm0:quad32) (va_x_xmm1:quad32)\n (va_x_xmm2:quad32) (va_x_xmm3:quad32) (va_x_xmm4:quad32) (va_x_xmm5:quad32) (va_x_xmm6:quad32)\n (va_x_xmm7:quad32) (va_x_xmm8:quad32) (va_x_xmm10:quad32) (va_x_heap5:vale_heap)\n (va_x_efl:Vale.X64.Flags.t) . let va_sM = va_upd_flags va_x_efl (va_upd_mem_heaplet 5\n va_x_heap5 (va_upd_xmm 10 va_x_xmm10 (va_upd_xmm 8 va_x_xmm8 (va_upd_xmm 7 va_x_xmm7\n (va_upd_xmm 6 va_x_xmm6 (va_upd_xmm 5 va_x_xmm5 (va_upd_xmm 4 va_x_xmm4 (va_upd_xmm 3 va_x_xmm3\n (va_upd_xmm 2 va_x_xmm2 (va_upd_xmm 1 va_x_xmm1 (va_upd_xmm 0 va_x_xmm0 (va_upd_reg64 rR11\n va_x_r11 (va_upd_reg64 rRcx va_x_rcx (va_upd_mem va_x_mem va_s0)))))))))))))) in va_get_ok\n va_sM /\\ (let (len:(va_int_range 1 1)) = 1 in Vale.X64.Decls.modifies_buffer128 inout_b\n (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5 va_sM) /\\ Vale.AES.GCTR.gctr_partial alg len\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b) key (va_get_xmm 11 va_s0) /\\ (let raw_quads =\n FStar.Seq.Base.append #quad32 completed_quads (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0)\n inout_b) in let input_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_quads) 0 total_bytes in let padded_bytes =\n Vale.AES.GCTR_s.pad_to_128_bits input_bytes in let input_quads =\n Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_bytes in l_and (FStar.Seq.Base.length\n #Vale.Def.Types_s.quad32 input_quads > 0) (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8\n va_sM) == Vale.AES.GHash.ghash_incremental h_LE old_hash input_quads))) ==> va_k va_sM (())))\n\n\nval va_wpProof_Gcm_extra_bytes : alg:algorithm -> inout_b:buffer128 -> key:(seq nat32) ->\n round_keys:(seq quad32) -> keys_b:buffer128 -> hkeys_b:buffer128 -> total_bytes:nat ->\n old_hash:quad32 -> completed_quads:(seq quad32) -> h_LE:quad32 -> va_s0:va_state ->\n va_k:(va_state -> unit -> Type0)\n -> Ghost (va_state & va_fuel & unit)\n (requires (va_t_require va_s0 /\\ va_wp_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b\n total_bytes old_hash completed_quads h_LE va_s0 va_k))\n (ensures (fun (va_sM, va_f0, va_g) -> va_t_ensure (va_code_Gcm_extra_bytes alg) ([va_Mod_flags;\n va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5;\n va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR11;\n va_Mod_reg64 rRcx; va_Mod_mem]) va_s0 va_k ((va_sM, va_f0, va_g))))\n[@\"opaque_to_smt\"]\nlet va_wpProof_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b total_bytes old_hash\n completed_quads h_LE va_s0 va_k =\n let (va_sM, va_f0) = va_lemma_Gcm_extra_bytes (va_code_Gcm_extra_bytes alg) va_s0 alg inout_b key\n round_keys keys_b hkeys_b total_bytes old_hash completed_quads h_LE in\n va_lemma_upd_update va_sM;\n assert (va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 5 va_sM (va_update_xmm 10\n va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5\n va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1\n va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR11 va_sM (va_update_reg64 rRcx va_sM\n (va_update_ok va_sM (va_update_mem va_sM va_s0)))))))))))))))));\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10; va_Mod_xmm 8; va_Mod_xmm\n 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1;\n va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_mem]) va_sM va_s0;\n let va_g = () in\n (va_sM, va_f0, va_g)\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_quick_Gcm_extra_bytes (alg:algorithm) (inout_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (total_bytes:nat) (old_hash:quad32)\n (completed_quads:(seq quad32)) (h_LE:quad32) : (va_quickCode unit (va_code_Gcm_extra_bytes alg)) =\n (va_QProc (va_code_Gcm_extra_bytes alg) ([va_Mod_flags; va_Mod_mem_heaplet 5; va_Mod_xmm 10;\n va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm\n 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR11; va_Mod_reg64 rRcx; va_Mod_mem])\n (va_wp_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b total_bytes old_hash\n completed_quads h_LE) (va_wpProof_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b\n total_bytes old_hash completed_quads h_LE))\n//--\n//-- Gcm_blocks128\n\nval va_code_Gcm_blocks128 : alg:algorithm -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_blocks128 alg =\n (va_Block (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRbx) (va_op_opr64_reg64 rRdi))\n (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR12) (va_op_opr64_reg64 rRdx)) (va_CCons\n (va_code_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRax)) (va_CCons\n (va_code_Ghash_buffer ()) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRdi)\n (va_op_opr64_reg64 rRbx)) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRdx)\n (va_op_opr64_reg64 rR12)) (va_CCons (va_code_Gctr_blocks128 alg) (va_CNil ())))))))))\n\nval va_codegen_success_Gcm_blocks128 : alg:algorithm -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_blocks128 alg =\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rRbx) (va_op_opr64_reg64 rRdi))\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR12) (va_op_opr64_reg64 rRdx))\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRax))\n (va_pbool_and (va_codegen_success_Ghash_buffer ()) (va_pbool_and (va_codegen_success_Mov64\n (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRbx)) (va_pbool_and (va_codegen_success_Mov64\n (va_op_dst_opr64_reg64 rRdx) (va_op_opr64_reg64 rR12)) (va_pbool_and\n (va_codegen_success_Gctr_blocks128 alg) (va_ttrue ()))))))))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_blocks128 (va_mods:va_mods_t) (alg:algorithm) (in_b:buffer128) (out_b:buffer128)\n (key:(seq nat32)) (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (h_LE:quad32)\n : (va_quickCode unit (va_code_Gcm_blocks128 alg)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 274 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRbx) (va_op_opr64_reg64 rRdi)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 275 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR12) (va_op_opr64_reg64 rRdx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 276 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRax)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 277 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Ghash_buffer hkeys_b in_b h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8\n va_old_s))) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 278 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRdi) (va_op_opr64_reg64 rRbx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 279 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRdx) (va_op_opr64_reg64 rR12)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 280 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gctr_blocks128 alg in_b out_b key round_keys keys_b) (va_QEmpty (()))))))))))\n\n\nval va_lemma_Gcm_blocks128 : va_b0:va_code -> va_s0:va_state -> alg:algorithm -> in_b:buffer128 ->\n out_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128 -> h_LE:quad32\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_blocks128 alg) va_s0 /\\ va_get_ok va_s0 /\\\n (sse_enabled /\\ Vale.X64.Decls.buffers_disjoint128 keys_b out_b /\\\n Vale.X64.Decls.buffers_disjoint128 hkeys_b out_b /\\ (Vale.X64.Decls.buffers_disjoint128 in_b\n out_b \\/ in_b == out_b) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0)\n (va_get_reg64 rRax va_s0) in_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1 va_s0) (va_get_reg64 rRdi va_s0) out_b\n (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\ va_get_reg64 rRax va_s0 + 16\n `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply`\n va_get_reg64 rRdx va_s0 < pow2_64 /\\ l_and (Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 in_b == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out_b)\n (Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b < pow2_32) /\\ va_get_reg64 rRdx\n va_s0 == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b /\\ va_get_xmm 9 va_s0 ==\n Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\\n va_get_reg64 rRdx va_s0 < pow2_32 /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rR8\n va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret)))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (Vale.X64.Decls.modifies_buffer128 out_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM) /\\ Vale.AES.GCTR.gctr_partial alg (va_get_reg64 rRdx va_s0) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b) key\n (va_get_xmm 11 va_s0) /\\ va_get_xmm 11 va_sM == Vale.AES.GCTR.inc32lite (va_get_xmm 11 va_s0)\n (va_get_reg64 rRdx va_s0) /\\ (va_get_reg64 rRdx va_s0 == 0 ==> l_and (va_get_xmm 8 va_sM ==\n va_get_xmm 8 va_s0) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b ==\n Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) out_b)) /\\ (va_get_reg64 rRdx va_s0 > 0 ==>\n l_and (va_get_reg64 rRdx va_s0 <= FStar.Seq.Base.length #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) ==> FStar.Seq.Base.length\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.slice #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) 0 (va_get_reg64 rRdx va_s0)) > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s0)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b)))) /\\ va_state_eq va_sM (va_update_flags va_sM\n (va_update_mem_heaplet 1 va_sM (va_update_xmm 10 va_sM (va_update_xmm 11 va_sM (va_update_xmm 8\n va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5 va_sM (va_update_xmm 4\n va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1 va_sM (va_update_xmm 0\n va_sM (va_update_reg64 rR12 va_sM (va_update_reg64 rRdx va_sM (va_update_reg64 rR10 va_sM\n (va_update_reg64 rR11 va_sM (va_update_reg64 rRdi va_sM (va_update_reg64 rRbx va_sM\n (va_update_ok va_sM (va_update_mem va_sM va_s0)))))))))))))))))))))))\n[@\"opaque_to_smt\"]\nlet va_lemma_Gcm_blocks128 va_b0 va_s0 alg in_b out_b key round_keys keys_b hkeys_b h_LE =\n let (va_mods:va_mods_t) = [va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11;\n va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm\n 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10;\n va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_ok; va_Mod_mem] in\n let va_qc = va_qcode_Gcm_blocks128 va_mods alg in_b out_b key round_keys keys_b hkeys_b h_LE in\n let (va_sM, va_fM, va_g) = va_wp_sound_code_norm (va_code_Gcm_blocks128 alg) va_qc va_s0 (fun\n va_s0 va_sM va_g -> let () = va_g in label va_range1\n \"***** POSTCONDITION NOT MET AT line 210 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\ (label va_range1\n \"***** POSTCONDITION NOT MET AT line 255 column 53 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 261 column 95 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.AES.GCTR.gctr_partial alg (va_get_reg64 rRdx va_s0) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b) key\n (va_get_xmm 11 va_s0)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 262 column 45 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_xmm 11 va_sM == Vale.AES.GCTR.inc32lite (va_get_xmm 11 va_s0) (va_get_reg64 rRdx\n va_s0)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 265 column 93 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRdx va_s0 == 0 ==> l_and (va_get_xmm 8 va_sM == va_get_xmm 8 va_s0)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b == Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) out_b)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 267 column 131 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRdx va_s0 > 0 ==> l_and (va_get_reg64 rRdx va_s0 <= FStar.Seq.Base.length\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) ==>\n FStar.Seq.Base.length #Vale.X64.Decls.quad32 (FStar.Seq.Base.slice #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) 0 (va_get_reg64 rRdx va_s0)) > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s0)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b))))) in\n assert_norm (va_qc.mods == va_mods);\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11; va_Mod_xmm\n 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2;\n va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10;\n va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_ok; va_Mod_mem]) va_sM va_s0;\n (va_sM, va_fM)\n\n[@ va_qattr]\nlet va_wp_Gcm_blocks128 (alg:algorithm) (in_b:buffer128) (out_b:buffer128) (key:(seq nat32))\n (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (h_LE:quad32) (va_s0:va_state)\n (va_k:(va_state -> unit -> Type0)) : Type0 =\n (va_get_ok va_s0 /\\ (sse_enabled /\\ Vale.X64.Decls.buffers_disjoint128 keys_b out_b /\\\n Vale.X64.Decls.buffers_disjoint128 hkeys_b out_b /\\ (Vale.X64.Decls.buffers_disjoint128 in_b\n out_b \\/ in_b == out_b) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0)\n (va_get_reg64 rRax va_s0) in_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1 va_s0) (va_get_reg64 rRdi va_s0) out_b\n (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0) Secret /\\ va_get_reg64 rRax va_s0 + 16\n `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply`\n va_get_reg64 rRdx va_s0 < pow2_64 /\\ l_and (Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 in_b == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out_b)\n (Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b < pow2_32) /\\ va_get_reg64 rRdx\n va_s0 == Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in_b /\\ va_get_xmm 9 va_s0 ==\n Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 202182159 134810123 67438087 66051 /\\\n va_get_reg64 rRdx va_s0 < pow2_32 /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rR8\n va_s0) (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 h_LE) /\\ Vale.X64.Decls.validSrcAddrs128\n (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0 - 32) hkeys_b 8 (va_get_mem_layout va_s0)\n Secret) /\\ (forall (va_x_mem:vale_heap) (va_x_rbx:nat64) (va_x_rdi:nat64) (va_x_r11:nat64)\n (va_x_r10:nat64) (va_x_rdx:nat64) (va_x_r12:nat64) (va_x_xmm0:quad32) (va_x_xmm1:quad32)\n (va_x_xmm2:quad32) (va_x_xmm3:quad32) (va_x_xmm4:quad32) (va_x_xmm5:quad32) (va_x_xmm6:quad32)\n (va_x_xmm7:quad32) (va_x_xmm8:quad32) (va_x_xmm11:quad32) (va_x_xmm10:quad32)\n (va_x_heap1:vale_heap) (va_x_efl:Vale.X64.Flags.t) . let va_sM = va_upd_flags va_x_efl\n (va_upd_mem_heaplet 1 va_x_heap1 (va_upd_xmm 10 va_x_xmm10 (va_upd_xmm 11 va_x_xmm11\n (va_upd_xmm 8 va_x_xmm8 (va_upd_xmm 7 va_x_xmm7 (va_upd_xmm 6 va_x_xmm6 (va_upd_xmm 5 va_x_xmm5\n (va_upd_xmm 4 va_x_xmm4 (va_upd_xmm 3 va_x_xmm3 (va_upd_xmm 2 va_x_xmm2 (va_upd_xmm 1 va_x_xmm1\n (va_upd_xmm 0 va_x_xmm0 (va_upd_reg64 rR12 va_x_r12 (va_upd_reg64 rRdx va_x_rdx (va_upd_reg64\n rR10 va_x_r10 (va_upd_reg64 rR11 va_x_r11 (va_upd_reg64 rRdi va_x_rdi (va_upd_reg64 rRbx\n va_x_rbx (va_upd_mem va_x_mem va_s0))))))))))))))))))) in va_get_ok va_sM /\\\n (Vale.X64.Decls.modifies_buffer128 out_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM) /\\ Vale.AES.GCTR.gctr_partial alg (va_get_reg64 rRdx va_s0) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b) key\n (va_get_xmm 11 va_s0) /\\ va_get_xmm 11 va_sM == Vale.AES.GCTR.inc32lite (va_get_xmm 11 va_s0)\n (va_get_reg64 rRdx va_s0) /\\ (va_get_reg64 rRdx va_s0 == 0 ==> l_and (va_get_xmm 8 va_sM ==\n va_get_xmm 8 va_s0) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out_b ==\n Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) out_b)) /\\ (va_get_reg64 rRdx va_s0 > 0 ==>\n l_and (va_get_reg64 rRdx va_s0 <= FStar.Seq.Base.length #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in_b) ==> FStar.Seq.Base.length\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.slice #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b) 0 (va_get_reg64 rRdx va_s0)) > 0)\n (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_sM) == Vale.AES.GHash.ghash_incremental\n h_LE (Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s0)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in_b)))) ==> va_k va_sM (())))\n\n\nval va_wpProof_Gcm_blocks128 : alg:algorithm -> in_b:buffer128 -> out_b:buffer128 -> key:(seq\n nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 -> hkeys_b:buffer128 -> h_LE:quad32 ->\n va_s0:va_state -> va_k:(va_state -> unit -> Type0)\n -> Ghost (va_state & va_fuel & unit)\n (requires (va_t_require va_s0 /\\ va_wp_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b\n h_LE va_s0 va_k))\n (ensures (fun (va_sM, va_f0, va_g) -> va_t_ensure (va_code_Gcm_blocks128 alg) ([va_Mod_flags;\n va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6;\n va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0;\n va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10; va_Mod_reg64 rR11; va_Mod_reg64 rRdi;\n va_Mod_reg64 rRbx; va_Mod_mem]) va_s0 va_k ((va_sM, va_f0, va_g))))\n[@\"opaque_to_smt\"]\nlet va_wpProof_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b h_LE va_s0 va_k =\n let (va_sM, va_f0) = va_lemma_Gcm_blocks128 (va_code_Gcm_blocks128 alg) va_s0 alg in_b out_b key\n round_keys keys_b hkeys_b h_LE in\n va_lemma_upd_update va_sM;\n assert (va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 1 va_sM (va_update_xmm 10\n va_sM (va_update_xmm 11 va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6\n va_sM (va_update_xmm 5 va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2\n va_sM (va_update_xmm 1 va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR12 va_sM\n (va_update_reg64 rRdx va_sM (va_update_reg64 rR10 va_sM (va_update_reg64 rR11 va_sM\n (va_update_reg64 rRdi va_sM (va_update_reg64 rRbx va_sM (va_update_ok va_sM (va_update_mem\n va_sM va_s0))))))))))))))))))))));\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10; va_Mod_xmm 11; va_Mod_xmm\n 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2;\n va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64 rR10;\n va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_mem]) va_sM va_s0;\n let va_g = () in\n (va_sM, va_f0, va_g)\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_quick_Gcm_blocks128 (alg:algorithm) (in_b:buffer128) (out_b:buffer128) (key:(seq nat32))\n (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (h_LE:quad32) : (va_quickCode\n unit (va_code_Gcm_blocks128 alg)) =\n (va_QProc (va_code_Gcm_blocks128 alg) ([va_Mod_flags; va_Mod_mem_heaplet 1; va_Mod_xmm 10;\n va_Mod_xmm 11; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm\n 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR12; va_Mod_reg64 rRdx; va_Mod_reg64\n rR10; va_Mod_reg64 rR11; va_Mod_reg64 rRdi; va_Mod_reg64 rRbx; va_Mod_mem])\n (va_wp_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b h_LE)\n (va_wpProof_Gcm_blocks128 alg in_b out_b key round_keys keys_b hkeys_b h_LE))\n//--\n//-- Gcm_blocks\n#push-options \"--z3rlimit 1000\"\n\nval va_code_Gcm_blocks : alg:algorithm -> offset:int -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_blocks alg offset =\n (va_Block (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR13) (va_op_opr64_reg64 rRcx))\n (va_CCons (va_code_AddLea64 (va_op_dst_opr64_reg64 rR9) (va_op_opr64_reg64 rR9) (va_const_opr64\n 32)) (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rRbx) (va_op_reg_opr64_reg64 rRsp)\n (offset + 0)) (va_CCons (va_code_Gcm_blocks_auth ()) (va_CCons (va_code_Load64_stack\n (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 8)) (va_CCons\n (va_code_Load64_stack (va_op_dst_opr64_reg64 rRsi) (va_op_reg_opr64_reg64 rRsp) (offset + 16))\n (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp)\n (offset + 24)) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rRcx) (va_op_opr64_reg64 rR13))\n (va_CCons (va_code_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (va_CCons\n (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 2) (va_op_xmm_xmm 1) (va_op_reg_opr64_reg64\n rR8) 0 Public) (va_CCons (va_code_Store128_buffer (va_op_heaplet_mem_heaplet 3)\n (va_op_reg_opr64_reg64 rRbp) (va_op_xmm_xmm 1) 0 Secret) (va_CCons (va_code_Load_one_lsb\n (va_op_xmm_xmm 10)) (va_CCons (va_code_VPaddd (va_op_xmm_xmm 1) (va_op_xmm_xmm 1)\n (va_op_xmm_xmm 10)) (va_CCons (va_code_AES_GCM_decrypt_6mult alg) (va_CCons\n (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 11) (va_op_reg_opr64_reg64\n rRbp) 32 Secret) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR8) (va_op_opr64_reg64 rRcx))\n (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp)\n (offset + 32)) (va_CCons (va_code_Load64_stack (va_op_dst_opr64_reg64 rRdi)\n (va_op_reg_opr64_reg64 rRsp) (offset + 40)) (va_CCons (va_code_Load64_stack\n (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp) (offset + 48)) (va_CCons\n (va_code_Mov64 (va_op_dst_opr64_reg64 rR14) (va_op_opr64_reg64 rRdx)) (va_CCons\n (va_code_InitPshufbMask (va_op_xmm_xmm 9) (va_op_reg_opr64_reg64 rR12)) (va_CCons\n (va_code_Pshufb (va_op_xmm_xmm 11) (va_op_xmm_xmm 9)) (va_CCons (va_code_Gcm_blocks128 alg)\n (va_CCons (va_code_Stack_lemma ()) (va_CCons (va_code_Add64 (va_op_dst_opr64_reg64 rR14)\n (va_opr_code_Stack (va_op_reg64_reg64 rRsp) (offset + 24) Public)) (va_CCons (va_code_IMul64\n (va_op_dst_opr64_reg64 rR14) (va_const_opr64 16)) (va_CCons (va_code_Load64_stack\n (va_op_dst_opr64_reg64 rR13) (va_op_reg_opr64_reg64 rRsp) (offset + 64)) (va_CCons (va_IfElse\n (va_cmp_gt (va_op_cmp_reg64 rR13) (va_op_cmp_reg64 rR14)) (va_Block (va_CCons\n (va_code_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp) (offset + 56))\n (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR10) (va_op_opr64_reg64 rR13)) (va_CCons\n (va_code_And64 (va_op_dst_opr64_reg64 rR10) (va_const_opr64 15)) (va_CCons\n (va_code_Gcm_extra_bytes alg) (va_CCons (va_Block (va_CNil ())) (va_CNil ()))))))) (va_Block\n (va_CNil ()))) (va_CCons (va_code_Mov64 (va_op_dst_opr64_reg64 rR11) (va_op_opr64_reg64 rR15))\n (va_CCons (va_code_Gcm_make_length_quad ()) (va_CCons (va_code_Ghash_register ()) (va_CCons\n (va_code_Load128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64\n rRbp) 0 Secret) (va_CCons (va_code_Gctr_register alg) (va_CCons (va_Block (va_CNil ()))\n (va_CNil ()))))))))))))))))))))))))))))))))))))\n\nval va_codegen_success_Gcm_blocks : alg:algorithm -> offset:int -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_blocks alg offset =\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR13) (va_op_opr64_reg64 rRcx))\n (va_pbool_and (va_codegen_success_AddLea64 (va_op_dst_opr64_reg64 rR9) (va_op_opr64_reg64 rR9)\n (va_const_opr64 32)) (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64\n rRbx) (va_op_reg_opr64_reg64 rRsp) (offset + 0)) (va_pbool_and\n (va_codegen_success_Gcm_blocks_auth ()) (va_pbool_and (va_codegen_success_Load64_stack\n (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 8)) (va_pbool_and\n (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRsi) (va_op_reg_opr64_reg64 rRsp)\n (offset + 16)) (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRdx)\n (va_op_reg_opr64_reg64 rRsp) (offset + 24)) (va_pbool_and (va_codegen_success_Mov64\n (va_op_dst_opr64_reg64 rRcx) (va_op_opr64_reg64 rR13)) (va_pbool_and (va_codegen_success_Mov128\n (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (va_pbool_and (va_codegen_success_Load128_buffer\n (va_op_heaplet_mem_heaplet 2) (va_op_xmm_xmm 1) (va_op_reg_opr64_reg64 rR8) 0 Public)\n (va_pbool_and (va_codegen_success_Store128_buffer (va_op_heaplet_mem_heaplet 3)\n (va_op_reg_opr64_reg64 rRbp) (va_op_xmm_xmm 1) 0 Secret) (va_pbool_and\n (va_codegen_success_Load_one_lsb (va_op_xmm_xmm 10)) (va_pbool_and (va_codegen_success_VPaddd\n (va_op_xmm_xmm 1) (va_op_xmm_xmm 1) (va_op_xmm_xmm 10)) (va_pbool_and\n (va_codegen_success_AES_GCM_decrypt_6mult alg) (va_pbool_and (va_codegen_success_Load128_buffer\n (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 11) (va_op_reg_opr64_reg64 rRbp) 32 Secret)\n (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR8) (va_op_opr64_reg64 rRcx))\n (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRax)\n (va_op_reg_opr64_reg64 rRsp) (offset + 32)) (va_pbool_and (va_codegen_success_Load64_stack\n (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 40)) (va_pbool_and\n (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp)\n (offset + 48)) (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR14)\n (va_op_opr64_reg64 rRdx)) (va_pbool_and (va_codegen_success_InitPshufbMask (va_op_xmm_xmm 9)\n (va_op_reg_opr64_reg64 rR12)) (va_pbool_and (va_codegen_success_Pshufb (va_op_xmm_xmm 11)\n (va_op_xmm_xmm 9)) (va_pbool_and (va_codegen_success_Gcm_blocks128 alg) (va_pbool_and\n (va_codegen_success_Stack_lemma ()) (va_pbool_and (va_codegen_success_Add64\n (va_op_dst_opr64_reg64 rR14) (va_opr_code_Stack (va_op_reg64_reg64 rRsp) (offset + 24) Public))\n (va_pbool_and (va_codegen_success_IMul64 (va_op_dst_opr64_reg64 rR14) (va_const_opr64 16))\n (va_pbool_and (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rR13)\n (va_op_reg_opr64_reg64 rRsp) (offset + 64)) (va_pbool_and (va_pbool_and\n (va_codegen_success_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp)\n (offset + 56)) (va_pbool_and (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR10)\n (va_op_opr64_reg64 rR13)) (va_pbool_and (va_codegen_success_And64 (va_op_dst_opr64_reg64 rR10)\n (va_const_opr64 15)) (va_codegen_success_Gcm_extra_bytes alg)))) (va_pbool_and\n (va_codegen_success_Mov64 (va_op_dst_opr64_reg64 rR11) (va_op_opr64_reg64 rR15)) (va_pbool_and\n (va_codegen_success_Gcm_make_length_quad ()) (va_pbool_and (va_codegen_success_Ghash_register\n ()) (va_pbool_and (va_codegen_success_Load128_buffer (va_op_heaplet_mem_heaplet 3)\n (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64 rRbp) 0 Secret) (va_pbool_and\n (va_codegen_success_Gctr_register alg) (va_ttrue ()))))))))))))))))))))))))))))))))))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_blocks (va_mods:va_mods_t) (alg:algorithm) (offset:int) (auth_b:buffer128)\n (abytes_b:buffer128) (in128x6_b:buffer128) (out128x6_b:buffer128) (in128_b:buffer128)\n (out128_b:buffer128) (inout_b:buffer128) (iv_b:buffer128) (scratch_b:buffer128) (key:(seq nat32))\n (round_keys:(seq quad32)) (keys_b:buffer128) (hkeys_b:buffer128) : (va_quickCode unit\n (va_code_Gcm_blocks alg offset)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in let\n (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 0) (va_get_stack va_s) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 8) (va_get_stack va_s) in let\n (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 16) (va_get_stack va_s) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 24) (va_get_stack va_s) in let\n (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 32) (va_get_stack va_s) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 40) (va_get_stack va_s) in let\n (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset\n + 48) (va_get_stack va_s) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 56) (va_get_stack va_s) in let\n (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s\n + offset + 64) (va_get_stack va_s) in let (h_LE:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.X64.Decls.buffer128_read hkeys_b 2\n (va_get_mem_heaplet 0 va_old_s)) in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 463 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR13) (va_op_opr64_reg64 rRcx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 464 column 13 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_AddLea64 (va_op_dst_opr64_reg64 rR9) (va_op_opr64_reg64 rR9) (va_const_opr64 32))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 465 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRbx) (va_op_reg_opr64_reg64 rRsp) (offset + 0))\n (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 466 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_blocks_auth auth_b abytes_b hkeys_b h_LE) (fun (va_s:va_state)\n (auth_quad_seq:(seq quad32)) -> let (y_0:quad32) = Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 0 0 0 0 in let (y_auth_bytes:quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s) in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 473 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 8))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 474 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRsi) (va_op_reg_opr64_reg64 rRsp) (offset + 16))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 475 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp) (offset + 24))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 476 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rRcx) (va_op_opr64_reg64 rR13)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 477 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov128 (va_op_xmm_xmm 0) (va_op_xmm_xmm 9)) (fun (va_s:va_state) _ -> let\n (iv_BE:Vale.X64.Decls.quad32) = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2\n va_old_s) in let (ctr_BE_1:quad32) = iv_BE in let (ctr_BE_2:quad32) = Vale.AES.GCTR_s.inc32\n iv_BE 1 in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 483 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 2) (va_op_xmm_xmm 1) (va_op_reg_opr64_reg64\n rR8) 0 Public iv_b 0) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 485 column 20 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Store128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_reg_opr64_reg64 rRbp)\n (va_op_xmm_xmm 1) 0 Secret scratch_b 0) (fun (va_s:va_state) _ -> let (j0:quad32) = va_get_xmm\n 1 va_s in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 487 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load_one_lsb (va_op_xmm_xmm 10)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 489 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_VPaddd (va_op_xmm_xmm 1) (va_op_xmm_xmm 1) (va_op_xmm_xmm 10)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 491 column 26 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_AES_GCM_decrypt_6mult alg h_LE iv_b in128x6_b out128x6_b scratch_b key round_keys\n keys_b hkeys_b) (fun (va_s:va_state) _ -> let (y_cipher128x6:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s) in let (auth_in:(seq quad32)) =\n auth_quad_seq in let (va_arg138:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b in let\n (va_arg137:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = auth_in in let\n (va_arg136:Vale.Def.Types_s.quad32) = y_auth_bytes in let (va_arg135:Vale.Def.Types_s.quad32) =\n y_0 in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 494 column 36 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GHash.lemma_ghash_incremental0_append h_LE va_arg135 va_arg136\n y_cipher128x6 va_arg137 va_arg138) (let auth_in = FStar.Seq.Base.append #quad32 auth_in\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b) in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 498 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 11)\n (va_op_reg_opr64_reg64 rRbp) 32 Secret scratch_b 2) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 499 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR8) (va_op_opr64_reg64 rRcx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 500 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp) (offset + 32))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 501 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRdi) (va_op_reg_opr64_reg64 rRsp) (offset + 40))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 502 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRdx) (va_op_reg_opr64_reg64 rRsp) (offset + 48))\n (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 503 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR14) (va_op_opr64_reg64 rRdx)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 504 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_InitPshufbMask (va_op_xmm_xmm 9) (va_op_reg_opr64_reg64 rR12)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 505 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Pshufb (va_op_xmm_xmm 11) (va_op_xmm_xmm 9)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 506 column 18 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_blocks128 alg in128_b out128_b key round_keys keys_b hkeys_b h_LE) (fun\n (va_s:va_state) _ -> let (y_cipher128:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s) in let (va_arg134:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b in let\n (va_arg133:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = auth_in in let\n (va_arg132:Vale.Def.Types_s.quad32) = y_0 in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 508 column 36 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GHash.lemma_ghash_incremental0_append h_LE va_arg132 y_cipher128x6\n y_cipher128 va_arg133 va_arg134) (let auth_in = FStar.Seq.Base.append #quad32 auth_in\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b) in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 512 column 22 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Stack_lemma (va_op_reg64_reg64 rRsp) (offset + 24) Public) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 512 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Add64 (va_op_dst_opr64_reg64 rR14) (va_opr_code_Stack (va_op_reg64_reg64 rRsp)\n (offset + 24) Public)) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 513 column 11 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_IMul64 (va_op_dst_opr64_reg64 rR14) (va_const_opr64 16)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 514 column 17 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rR13) (va_op_reg_opr64_reg64 rRsp) (offset + 64))\n (fun (va_s:va_state) _ -> let (y_inout:Vale.Def.Types_s.quad32) = y_cipher128 in let\n (plain_byte_seq:(seq quad32)) = empty_seq_quad32 in let (cipher_byte_seq:(seq quad32)) =\n empty_seq_quad32 in let (va_arg131:Vale.Def.Types_s.quad32) = va_get_xmm 11 va_s in let\n (va_arg130:(FStar.Seq.Base.seq Vale.Def.Types_s.nat32)) = key in let\n (va_arg129:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = cipher_byte_seq in let\n (va_arg128:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = plain_byte_seq in let\n (va_arg127:Vale.AES.AES_common_s.algorithm) = alg in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 519 column 29 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.gctr_partial_opaque_init va_arg127 va_arg128 va_arg129 va_arg130\n va_arg131) (let (total_bytes:(va_int_at_least 0)) = FStar.Seq.Base.length #quad32 auth_quad_seq\n `op_Multiply` 16 + plain_num_bytes in va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 523 column 8 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_qIf va_mods (Cmp_gt (va_op_cmp_reg64 rR13) (va_op_cmp_reg64 rR14)) (qblock va_mods (fun\n (va_s:va_state) -> va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 525 column 21 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load64_stack (va_op_dst_opr64_reg64 rRax) (va_op_reg_opr64_reg64 rRsp) (offset + 56))\n (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 526 column 14 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR10) (va_op_opr64_reg64 rR13)) (fun (va_s:va_state) _\n -> va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 527 column 26 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.Poly1305.Math.lemma_poly_bits64 ()) (va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 528 column 14 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_And64 (va_op_dst_opr64_reg64 rR10) (va_const_opr64 15)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 532 column 24 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_extra_bytes alg inout_b key round_keys keys_b hkeys_b total_bytes y_0 auth_in\n h_LE) (fun (va_s:va_state) _ -> let y_inout = Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm\n 8 va_s) in let (raw_auth_quads:(FStar.Seq.Base.seq quad32)) = FStar.Seq.Base.append #quad32\n auth_in (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_old_s) inout_b) in va_qAssertSquash\n va_range1\n \"***** EXPRESSION PRECONDITIONS NOT MET WITHIN line 536 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n ((fun a_1906 (s_1907:(FStar.Seq.Base.seq a_1906)) (i_1908:Prims.nat) (j_1909:Prims.nat) -> let\n (j_1869:Prims.nat) = j_1909 in Prims.b2t (Prims.op_AmpAmp (Prims.op_LessThanOrEqual i_1908\n j_1869) (Prims.op_LessThanOrEqual j_1869 (FStar.Seq.Base.length #a_1906 s_1907))))\n Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_auth_quads) 0 total_bytes)\n (fun _ -> let (auth_input_bytes:(FStar.Seq.Base.seq Vale.Def.Types_s.nat8)) =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n raw_auth_quads) 0 total_bytes in let (padded_auth_bytes:(FStar.Seq.Base.seq\n Vale.Def.Types_s.nat8)) = Vale.AES.GCTR_s.pad_to_128_bits auth_input_bytes in let auth_in =\n Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_auth_bytes in let plain_byte_seq =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_old_s) inout_b in let cipher_byte_seq =\n Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s) inout_b in va_QEmpty ((auth_in,\n cipher_byte_seq, plain_byte_seq, y_inout)))))))))) (qblock va_mods (fun (va_s:va_state) ->\n va_QEmpty ((auth_in, cipher_byte_seq, plain_byte_seq, y_inout))))) (fun (va_s:va_state) va_g ->\n let ((auth_in:(seq quad32)), (cipher_byte_seq:(seq quad32)), (plain_byte_seq:(seq quad32)),\n (y_inout:Vale.Def.Types_s.quad32)) = va_g in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 547 column 10 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Mov64 (va_op_dst_opr64_reg64 rR11) (va_op_opr64_reg64 rR15)) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 548 column 25 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_make_length_quad ()) (fun (va_s:va_state) _ -> let\n (length_quad32:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 0\n va_s) in va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 551 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Ghash_register hkeys_b h_LE y_inout) (fun (va_s:va_state) _ -> let\n (y_final:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32 (va_get_xmm 8 va_s)\n in va_QSeq va_range1\n \"***** PRECONDITION NOT MET AT line 554 column 19 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Load128_buffer (va_op_heaplet_mem_heaplet 3) (va_op_xmm_xmm 0) (va_op_reg_opr64_reg64\n rRbp) 0 Secret scratch_b 0) (va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 557 column 18 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gctr_register alg key round_keys keys_b) (fun (va_s:va_state) _ -> let\n (va_arg126:Vale.Def.Types_s.quad32) = va_get_xmm 8 va_s in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 560 column 40 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.Arch.Types.le_seq_quad32_to_bytes_of_singleton va_arg126)\n (va_qAssertSquash va_range1\n \"***** EXPRESSION PRECONDITIONS NOT MET WITHIN line 561 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n ((fun (icb_BE_677:Vale.Def.Types_s.quad32) (plain_LE_678:Vale.Def.Types_s.quad32)\n (alg_679:Vale.AES.AES_common_s.algorithm) (key_680:(FStar.Seq.Base.seq Vale.Def.Types_s.nat32))\n (i_681:Prims.int) -> Vale.AES.AES_s.is_aes_key_LE alg_679 key_680) j0 y_final alg key 0) (fun _\n -> va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 561 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_xmm 8 va_s == Vale.AES.GCTR_s.gctr_encrypt_block j0 y_final alg key 0) (let\n (plain128:(FStar.Seq.Base.seq Vale.X64.Decls.quad32)) = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b) in let\n (cipher128:(FStar.Seq.Base.seq Vale.X64.Decls.quad32)) = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) in128_b) in va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 566 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (FStar.Seq.Base.length #quad32 plain_byte_seq == 0 ==> FStar.Seq.Base.equal\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 plain128 plain_byte_seq)\n plain128) (va_qAssert va_range1\n \"***** PRECONDITION NOT MET AT line 567 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (FStar.Seq.Base.length #quad32 cipher_byte_seq == 0 ==> FStar.Seq.Base.equal\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 cipher128 cipher_byte_seq)\n cipher128) (let (va_arg125:Vale.Def.Types_s.quad32) = Vale.AES.GCTR.inc32lite ctr_BE_2 len128x6\n in let (va_arg124:Vale.Def.Types_s.quad32) = ctr_BE_2 in let (va_arg123:(FStar.Seq.Base.seq\n Vale.Def.Types_s.nat32)) = key in let (va_arg122:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32))\n = Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) out128_b in let\n (va_arg121:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_old_s) in128_b in let (va_arg120:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s) out128x6_b in let\n (va_arg119:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_old_s) in128x6_b in let (va_arg118:Prims.nat) = len128 in let\n (va_arg117:Prims.nat) = len128x6 in let (va_arg116:Vale.AES.AES_common_s.algorithm) = alg in\n va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 569 column 30 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.lemma_gctr_partial_append va_arg116 va_arg117 va_arg118\n va_arg119 va_arg120 va_arg121 va_arg122 va_arg123 va_arg124 va_arg125) (let\n (va_arg115:Vale.Def.Types_s.quad32) = Vale.AES.GCTR.inc32lite (Vale.AES.GCTR.inc32lite ctr_BE_2\n len128x6) len128 in let (va_arg114:Vale.Def.Types_s.quad32) = ctr_BE_2 in let\n (va_arg113:(FStar.Seq.Base.seq Vale.Def.Types_s.nat32)) = key in let\n (va_arg112:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = cipher_byte_seq in let\n (va_arg111:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = plain_byte_seq in let\n (va_arg110:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s) out128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) out128_b) in let\n (va_arg109:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) in128_b) in let (va_arg108:Prims.nat) =\n FStar.Seq.Base.length #quad32 plain_byte_seq in let (va_arg107:Prims.nat) = len128x6 + len128\n in let (va_arg106:Vale.AES.AES_common_s.algorithm) = alg in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 575 column 30 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCTR.lemma_gctr_partial_append va_arg106 va_arg107 va_arg108\n va_arg109 va_arg110 va_arg111 va_arg112 va_arg113 va_arg114 va_arg115) (let\n (va_arg105:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = auth_in in let\n (va_arg104:Vale.Def.Types_s.quad32) = y_0 in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 583 column 23 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GHash.lemma_hash_append2 h_LE va_arg104 y_inout y_final va_arg105\n length_quad32) (let auth_in = FStar.Seq.Base.append #quad32 auth_in (FStar.Seq.Base.create\n #Vale.Def.Types_s.quad32 1 length_quad32) in let (va_arg103:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = auth_in in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 585 column 31 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GHash.ghash_incremental_to_ghash h_LE va_arg103) (va_QEmpty\n (()))))))))))))))))))))))))))))))))))))))))))))))))\n\n\nval va_lemma_Gcm_blocks : va_b0:va_code -> va_s0:va_state -> alg:algorithm -> offset:int ->\n auth_b:buffer128 -> abytes_b:buffer128 -> in128x6_b:buffer128 -> out128x6_b:buffer128 ->\n in128_b:buffer128 -> out128_b:buffer128 -> inout_b:buffer128 -> iv_b:buffer128 ->\n scratch_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) -> keys_b:buffer128 ->\n hkeys_b:buffer128\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_blocks alg offset) va_s0 /\\ va_get_ok va_s0 /\\\n (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in let (h_LE:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.X64.Decls.buffer128_read hkeys_b 2\n (va_get_mem_heaplet 0 va_s0)) in sse_enabled /\\ movbe_enabled /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0)\n Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp\n va_s0 + offset + 8) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 24) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 40) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 56) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet\n 1 va_s0) (va_get_reg64 rRdi va_s0) auth_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0)\n Secret /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 7 va_s0) abytes_ptr abytes_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 2\n va_s0) (va_get_reg64 rR8 va_s0) iv_b 1 (va_get_mem_layout va_s0) Public /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 6 va_s0) in128x6_ptr in128x6_b len128x6\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 6\n va_s0) out128x6_ptr out128x6_b len128x6 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0) in128_ptr in128_b len128\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1\n va_s0) out128_ptr out128_b len128 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) inout_ptr inout_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 3\n va_s0) (va_get_reg64 rRbp va_s0) scratch_b 9 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0) hkeys_b 8\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.buffer_disjoints128 iv_b ([keys_b;\n scratch_b; in128x6_b; out128x6_b; hkeys_b; in128_b; out128_b; inout_b]) /\\\n Vale.X64.Decls.buffer_disjoints128 scratch_b ([keys_b; in128x6_b; out128x6_b; in128_b;\n out128_b; inout_b; hkeys_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128x6_b ([keys_b;\n hkeys_b; in128_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128_b ([keys_b; hkeys_b;\n out128x6_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 inout_b ([keys_b; hkeys_b;\n out128x6_b; out128_b]) /\\ (Vale.X64.Decls.buffers_disjoint128 in128x6_b out128x6_b \\/ in128x6_b\n == out128x6_b) /\\ (Vale.X64.Decls.buffers_disjoint128 in128_b out128_b \\/ in128_b == out128_b)\n /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ in128x6_ptr\n + 16 `op_Multiply` len128x6 < pow2_64 /\\ out128x6_ptr + 16 `op_Multiply` len128x6 < pow2_64 /\\\n in128_ptr + 16 `op_Multiply` len128 < pow2_64 /\\ out128_ptr + 16 `op_Multiply` len128 < pow2_64\n /\\ inout_ptr + 16 < pow2_64 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 auth_b ==\n va_get_reg64 rRdx va_s0 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 abytes_b == 1\n /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128x6_b ==\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out128x6_b /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 out128_b /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128\n in128x6_b == len128x6 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b ==\n len128 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 inout_b == 1 /\\\n plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32 /\\ va_get_reg64 rR9 va_s0 + 32 <\n pow2_64 /\\ Vale.X64.Memory.buffer_addr #Vale.X64.Memory.vuint128 keys_b (va_get_mem_heaplet 0\n va_s0) + 128 < pow2_64 /\\ len128x6 `op_Modulus` 6 == 0 /\\ (len128x6 > 0 ==> len128x6 >= 6) /\\\n 12 + len128x6 + 6 < pow2_32 /\\ (va_mul_nat len128x6 (128 `op_Division` 8) + va_mul_nat len128\n (128 `op_Division` 8) <= plain_num_bytes /\\ plain_num_bytes < va_mul_nat len128x6 (128\n `op_Division` 8) + va_mul_nat len128 (128 `op_Division` 8) + 128 `op_Division` 8) /\\\n (va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) <= va_get_reg64 rRsi va_s0 /\\\n va_get_reg64 rRsi va_s0 < va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) + 128\n `op_Division` 8) /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rRcx va_s0)\n (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 (Vale.AES.AES_s.aes_encrypt_LE alg key\n (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0))))))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in let (h_LE:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.X64.Decls.buffer128_read hkeys_b 2\n (va_get_mem_heaplet 0 va_s0)) in Vale.X64.Decls.modifies_buffer128 out128_b (va_get_mem_heaplet\n 1 va_s0) (va_get_mem_heaplet 1 va_sM) /\\ Vale.X64.Decls.modifies_buffer128 iv_b\n (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2 va_sM) /\\ Vale.X64.Decls.modifies_buffer128\n scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3 va_sM) /\\\n Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM) /\\ Vale.X64.Decls.modifies_buffer128 out128x6_b (va_get_mem_heaplet 6 va_s0)\n (va_get_mem_heaplet 6 va_sM) /\\ plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32\n /\\ (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in let\n (ctr_BE_1:quad32) = iv_BE in let (ctr_BE_2:quad32) = Vale.AES.GCTR_s.inc32 iv_BE 1 in let\n (plain_in:(seq quad32)) = (if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) then FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b) else FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in128_b)) in let (cipher_out:(seq quad32)) = (if (plain_num_bytes\n > (len128x6 + len128) `op_Multiply` 128 `op_Division` 8) then FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM)\n out128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM) inout_b) else\n FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM)\n out128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out128_b)) in let\n (cipher_bound:nat) = (if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128 `op_Division`\n 8) then (len128x6 + len128 + 1) else (len128x6 + len128)) in Vale.AES.GCTR.gctr_partial alg\n cipher_bound plain_in cipher_out key ctr_BE_2 /\\ (let (length_quad:quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.Def.Types_s.insert_nat64\n (Vale.Def.Types_s.insert_nat64 (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0) (8\n `op_Multiply` va_get_reg64 rRsi va_s0) 1) (8 `op_Multiply` plain_num_bytes) 0) in let\n (raw_auth_quads:(seq quad32)) = (if (va_get_reg64 rRsi va_s0 > va_get_reg64 rRdx va_s0\n `op_Multiply` 128 `op_Division` 8) then FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 7 va_s0) abytes_b) else Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0)\n auth_b) in let (auth_input_bytes:(seq nat8)) = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_auth_quads) 0 (va_get_reg64 rRsi va_s0) in let\n (padded_auth_bytes:(seq nat8)) = Vale.AES.GCTR_s.pad_to_128_bits auth_input_bytes in let\n (auth_quad_seq:(seq quad32)) = Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_auth_bytes in let\n (raw_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32 (FStar.Seq.Base.append #quad32\n auth_quad_seq (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b))\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b) in let (total_bytes:nat) =\n FStar.Seq.Base.length #quad32 auth_quad_seq `op_Multiply` 16 + plain_num_bytes in let\n (raw_quad_seq:(seq quad32)) = (if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) then (let (ab:(seq nat8)) = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes (FStar.Seq.Base.append #quad32 raw_quad_seq\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b))) 0 total_bytes in let (pb:(seq\n nat8)) = Vale.AES.GCTR_s.pad_to_128_bits ab in Vale.Def.Types_s.le_bytes_to_seq_quad32 pb) else\n raw_quad_seq) in let (auth_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32 raw_quad_seq\n (FStar.Seq.Base.create #quad32 1 length_quad) in va_get_xmm 8 va_sM ==\n Vale.AES.GCTR_s.gctr_encrypt_block ctr_BE_1 (Vale.AES.GHash_s.ghash_LE h_LE auth_quad_seq) alg\n key 0))) /\\ va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 6 va_sM\n (va_update_mem_heaplet 5 va_sM (va_update_mem_heaplet 3 va_sM (va_update_mem_heaplet 2 va_sM\n (va_update_mem_heaplet 1 va_sM (va_update_xmm 15 va_sM (va_update_xmm 14 va_sM (va_update_xmm\n 13 va_sM (va_update_xmm 12 va_sM (va_update_xmm 11 va_sM (va_update_xmm 10 va_sM (va_update_xmm\n 9 va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5\n va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1\n va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR15 va_sM (va_update_reg64 rR14 va_sM\n (va_update_reg64 rR13 va_sM (va_update_reg64 rR12 va_sM (va_update_reg64 rR11 va_sM\n (va_update_reg64 rR10 va_sM (va_update_reg64 rR9 va_sM (va_update_reg64 rR8 va_sM\n (va_update_reg64 rRbp va_sM (va_update_reg64 rRsi va_sM (va_update_reg64 rRdi va_sM\n (va_update_reg64 rRdx va_sM (va_update_reg64 rRcx va_sM (va_update_reg64 rRbx va_sM\n (va_update_reg64 rRax va_sM (va_update_ok va_sM (va_update_mem va_sM\n va_s0)))))))))))))))))))))))))))))))))))))))))\n[@\"opaque_to_smt\"]\nlet va_lemma_Gcm_blocks va_b0 va_s0 alg offset auth_b abytes_b in128x6_b out128x6_b in128_b\n out128_b inout_b iv_b scratch_b key round_keys keys_b hkeys_b =\n let (va_mods:va_mods_t) = [va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5;\n va_Mod_mem_heaplet 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14;\n va_Mod_xmm 13; va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8;\n va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm\n 1; va_Mod_xmm 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12;\n va_Mod_reg64 rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp;\n va_Mod_reg64 rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx;\n va_Mod_reg64 rRax; va_Mod_ok; va_Mod_mem] in\n let va_qc = va_qcode_Gcm_blocks va_mods alg offset auth_b abytes_b in128x6_b out128x6_b in128_b\n out128_b inout_b iv_b scratch_b key round_keys keys_b hkeys_b in\n let (va_sM, va_fM, va_g) = va_wp_sound_code_norm (va_code_Gcm_blocks alg offset) va_qc va_s0 (fun\n va_s0 va_sM va_g -> let () = va_g in label va_range1\n \"***** POSTCONDITION NOT MET AT line 283 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\ (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64\n (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0) in let\n (in128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 8) (va_get_stack va_s0) in let (out128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack va_s0) in\n let (len128x6:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 24) (va_get_stack va_s0) in let (in128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack va_s0) in\n let (out128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 40) (va_get_stack va_s0) in let (len128:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack va_s0) in\n let (inout_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 56) (va_get_stack va_s0) in let (plain_num_bytes:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack va_s0) in\n let (h_LE:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.X64.Decls.buffer128_read hkeys_b 2 (va_get_mem_heaplet 0 va_s0)) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 396 column 56 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out128_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 397 column 52 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 iv_b (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 398 column 57 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 399 column 55 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 400 column 58 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out128x6_b (va_get_mem_heaplet 6 va_s0) (va_get_mem_heaplet\n 6 va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 403 column 39 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (plain_num_bytes < pow2_32) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 404 column 38 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRsi va_s0 < pow2_32) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 406 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 408 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (ctr_BE_1:quad32) = iv_BE in label va_range1\n \"***** POSTCONDITION NOT MET AT line 409 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (ctr_BE_2:quad32) = Vale.AES.GCTR_s.inc32 iv_BE 1 in label va_range1\n \"***** POSTCONDITION NOT MET AT line 412 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (plain_in:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b)) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in128_b)) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 421 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (cipher_out:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM) out128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b)) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_sM) out128_b)) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 430 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (cipher_bound:nat) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> len128x6 + len128 + 1) (fun _ -> len128x6 + len128) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 434 column 77 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.AES.GCTR.gctr_partial alg cipher_bound plain_in cipher_out key ctr_BE_2) /\\ label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 438 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (length_quad:quad32) = Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.Def.Types_s.insert_nat64 (Vale.Def.Types_s.insert_nat64 (Vale.Def.Words_s.Mkfour\n #Vale.Def.Types_s.nat32 0 0 0 0) (8 `op_Multiply` va_get_reg64 rRsi va_s0) 1) (8 `op_Multiply`\n plain_num_bytes) 0) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 440 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (raw_auth_quads:(seq quad32)) = va_if (va_get_reg64 rRsi va_s0 > va_get_reg64 rRdx va_s0\n `op_Multiply` 128 `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 7 va_s0) abytes_b)) (fun _ -> Vale.X64.Decls.s128 (va_get_mem_heaplet 1\n va_s0) auth_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 444 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (auth_input_bytes:(seq nat8)) = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_auth_quads) 0 (va_get_reg64 rRsi va_s0) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 445 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (padded_auth_bytes:(seq nat8)) = Vale.AES.GCTR_s.pad_to_128_bits auth_input_bytes in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 446 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (auth_quad_seq:(seq quad32)) = Vale.Def.Types_s.le_bytes_to_seq_quad32 padded_auth_bytes\n in label va_range1\n \"***** POSTCONDITION NOT MET AT line 448 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (raw_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32 (FStar.Seq.Base.append #quad32\n auth_quad_seq (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b))\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 452 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (total_bytes:nat) = FStar.Seq.Base.length #quad32 auth_quad_seq `op_Multiply` 16 +\n plain_num_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 453 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (raw_quad_seq:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply`\n 128 `op_Division` 8) (fun _ -> let (ab:(seq nat8)) = FStar.Seq.Base.slice\n #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes (FStar.Seq.Base.append #quad32\n raw_quad_seq (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b))) 0 total_bytes in let\n (pb:(seq nat8)) = Vale.AES.GCTR_s.pad_to_128_bits ab in Vale.Def.Types_s.le_bytes_to_seq_quad32\n pb) (fun _ -> raw_quad_seq) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 460 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let (auth_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32 raw_quad_seq\n (FStar.Seq.Base.create #quad32 1 length_quad) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 461 column 106 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_xmm 8 va_sM == Vale.AES.GCTR_s.gctr_encrypt_block ctr_BE_1 (Vale.AES.GHash_s.ghash_LE\n h_LE auth_quad_seq) alg key 0)))))))))))))))))) in\n assert_norm (va_qc.mods == va_mods);\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet\n 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13;\n va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm\n 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64\n rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64\n rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64\n rRax; va_Mod_ok; va_Mod_mem]) va_sM va_s0;\n (va_sM, va_fM)\n\n[@ va_qattr]\nlet va_wp_Gcm_blocks (alg:algorithm) (offset:int) (auth_b:buffer128) (abytes_b:buffer128)\n (in128x6_b:buffer128) (out128x6_b:buffer128) (in128_b:buffer128) (out128_b:buffer128)\n (inout_b:buffer128) (iv_b:buffer128) (scratch_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) (va_s0:va_state) (va_k:(va_state -> unit ->\n Type0)) : Type0 =\n (va_get_ok va_s0 /\\ (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64\n (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0) in let\n (in128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 8) (va_get_stack va_s0) in let (out128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack va_s0) in\n let (len128x6:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 24) (va_get_stack va_s0) in let (in128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack va_s0) in\n let (out128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 40) (va_get_stack va_s0) in let (len128:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack va_s0) in\n let (inout_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 56) (va_get_stack va_s0) in let (plain_num_bytes:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack va_s0) in\n let (h_LE:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.X64.Decls.buffer128_read hkeys_b 2 (va_get_mem_heaplet 0 va_s0)) in sse_enabled /\\\n movbe_enabled /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 0)\n (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64\n (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 24) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 40) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 56) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet\n 1 va_s0) (va_get_reg64 rRdi va_s0) auth_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0)\n Secret /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 7 va_s0) abytes_ptr abytes_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 2\n va_s0) (va_get_reg64 rR8 va_s0) iv_b 1 (va_get_mem_layout va_s0) Public /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 6 va_s0) in128x6_ptr in128x6_b len128x6\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 6\n va_s0) out128x6_ptr out128x6_b len128x6 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0) in128_ptr in128_b len128\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1\n va_s0) out128_ptr out128_b len128 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) inout_ptr inout_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 3\n va_s0) (va_get_reg64 rRbp va_s0) scratch_b 9 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0) hkeys_b 8\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.buffer_disjoints128 iv_b ([keys_b;\n scratch_b; in128x6_b; out128x6_b; hkeys_b; in128_b; out128_b; inout_b]) /\\\n Vale.X64.Decls.buffer_disjoints128 scratch_b ([keys_b; in128x6_b; out128x6_b; in128_b;\n out128_b; inout_b; hkeys_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128x6_b ([keys_b;\n hkeys_b; in128_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128_b ([keys_b; hkeys_b;\n out128x6_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 inout_b ([keys_b; hkeys_b;\n out128x6_b; out128_b]) /\\ (Vale.X64.Decls.buffers_disjoint128 in128x6_b out128x6_b \\/ in128x6_b\n == out128x6_b) /\\ (Vale.X64.Decls.buffers_disjoint128 in128_b out128_b \\/ in128_b == out128_b)\n /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ in128x6_ptr\n + 16 `op_Multiply` len128x6 < pow2_64 /\\ out128x6_ptr + 16 `op_Multiply` len128x6 < pow2_64 /\\\n in128_ptr + 16 `op_Multiply` len128 < pow2_64 /\\ out128_ptr + 16 `op_Multiply` len128 < pow2_64\n /\\ inout_ptr + 16 < pow2_64 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 auth_b ==\n va_get_reg64 rRdx va_s0 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 abytes_b == 1\n /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128x6_b ==\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out128x6_b /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 out128_b /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128\n in128x6_b == len128x6 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b ==\n len128 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 inout_b == 1 /\\\n plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32 /\\ va_get_reg64 rR9 va_s0 + 32 <\n pow2_64 /\\ Vale.X64.Memory.buffer_addr #Vale.X64.Memory.vuint128 keys_b (va_get_mem_heaplet 0\n va_s0) + 128 < pow2_64 /\\ len128x6 `op_Modulus` 6 == 0 /\\ (len128x6 > 0 ==> len128x6 >= 6) /\\\n 12 + len128x6 + 6 < pow2_32 /\\ (va_mul_nat len128x6 (128 `op_Division` 8) + va_mul_nat len128\n (128 `op_Division` 8) <= plain_num_bytes /\\ plain_num_bytes < va_mul_nat len128x6 (128\n `op_Division` 8) + va_mul_nat len128 (128 `op_Division` 8) + 128 `op_Division` 8) /\\\n (va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) <= va_get_reg64 rRsi va_s0 /\\\n va_get_reg64 rRsi va_s0 < va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) + 128\n `op_Division` 8) /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rRcx va_s0)\n (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 (Vale.AES.AES_s.aes_encrypt_LE alg key\n (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0)))) /\\ (forall (va_x_mem:vale_heap)\n (va_x_rax:nat64) (va_x_rbx:nat64) (va_x_rcx:nat64) (va_x_rdx:nat64) (va_x_rdi:nat64)\n (va_x_rsi:nat64) (va_x_rbp:nat64) (va_x_r8:nat64) (va_x_r9:nat64) (va_x_r10:nat64)\n (va_x_r11:nat64) (va_x_r12:nat64) (va_x_r13:nat64) (va_x_r14:nat64) (va_x_r15:nat64)\n (va_x_xmm0:quad32) (va_x_xmm1:quad32) (va_x_xmm2:quad32) (va_x_xmm3:quad32) (va_x_xmm4:quad32)\n (va_x_xmm5:quad32) (va_x_xmm6:quad32) (va_x_xmm7:quad32) (va_x_xmm8:quad32) (va_x_xmm9:quad32)\n (va_x_xmm10:quad32) (va_x_xmm11:quad32) (va_x_xmm12:quad32) (va_x_xmm13:quad32)\n (va_x_xmm14:quad32) (va_x_xmm15:quad32) (va_x_heap1:vale_heap) (va_x_heap2:vale_heap)\n (va_x_heap3:vale_heap) (va_x_heap5:vale_heap) (va_x_heap6:vale_heap)\n (va_x_efl:Vale.X64.Flags.t) . let va_sM = va_upd_flags va_x_efl (va_upd_mem_heaplet 6\n va_x_heap6 (va_upd_mem_heaplet 5 va_x_heap5 (va_upd_mem_heaplet 3 va_x_heap3\n (va_upd_mem_heaplet 2 va_x_heap2 (va_upd_mem_heaplet 1 va_x_heap1 (va_upd_xmm 15 va_x_xmm15\n (va_upd_xmm 14 va_x_xmm14 (va_upd_xmm 13 va_x_xmm13 (va_upd_xmm 12 va_x_xmm12 (va_upd_xmm 11\n va_x_xmm11 (va_upd_xmm 10 va_x_xmm10 (va_upd_xmm 9 va_x_xmm9 (va_upd_xmm 8 va_x_xmm8\n (va_upd_xmm 7 va_x_xmm7 (va_upd_xmm 6 va_x_xmm6 (va_upd_xmm 5 va_x_xmm5 (va_upd_xmm 4 va_x_xmm4\n (va_upd_xmm 3 va_x_xmm3 (va_upd_xmm 2 va_x_xmm2 (va_upd_xmm 1 va_x_xmm1 (va_upd_xmm 0 va_x_xmm0\n (va_upd_reg64 rR15 va_x_r15 (va_upd_reg64 rR14 va_x_r14 (va_upd_reg64 rR13 va_x_r13\n (va_upd_reg64 rR12 va_x_r12 (va_upd_reg64 rR11 va_x_r11 (va_upd_reg64 rR10 va_x_r10\n (va_upd_reg64 rR9 va_x_r9 (va_upd_reg64 rR8 va_x_r8 (va_upd_reg64 rRbp va_x_rbp (va_upd_reg64\n rRsi va_x_rsi (va_upd_reg64 rRdi va_x_rdi (va_upd_reg64 rRdx va_x_rdx (va_upd_reg64 rRcx\n va_x_rcx (va_upd_reg64 rRbx va_x_rbx (va_upd_reg64 rRax va_x_rax (va_upd_mem va_x_mem\n va_s0))))))))))))))))))))))))))))))))))))) in va_get_ok va_sM /\\ (let\n (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in let (h_LE:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.X64.Decls.buffer128_read hkeys_b 2\n (va_get_mem_heaplet 0 va_s0)) in Vale.X64.Decls.modifies_buffer128 out128_b (va_get_mem_heaplet\n 1 va_s0) (va_get_mem_heaplet 1 va_sM) /\\ Vale.X64.Decls.modifies_buffer128 iv_b\n (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2 va_sM) /\\ Vale.X64.Decls.modifies_buffer128\n scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3 va_sM) /\\\n Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM) /\\ Vale.X64.Decls.modifies_buffer128 out128x6_b (va_get_mem_heaplet 6 va_s0)\n (va_get_mem_heaplet 6 va_sM) /\\ plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32\n /\\ (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in let\n (ctr_BE_1:quad32) = iv_BE in let (ctr_BE_2:quad32) = Vale.AES.GCTR_s.inc32 iv_BE 1 in let\n (plain_in:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b)) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) in128_b)) in let (cipher_out:(seq quad32)) = va_if\n (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128 `op_Division` 8) (fun _ ->\n FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_sM) out128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM)\n inout_b)) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM)\n out128_b)) in let (cipher_bound:nat) = va_if (plain_num_bytes > (len128x6 + len128)\n `op_Multiply` 128 `op_Division` 8) (fun _ -> len128x6 + len128 + 1) (fun _ -> len128x6 +\n len128) in Vale.AES.GCTR.gctr_partial alg cipher_bound plain_in cipher_out key ctr_BE_2 /\\ (let\n (length_quad:quad32) = Vale.Def.Types_s.reverse_bytes_quad32 (Vale.Def.Types_s.insert_nat64\n (Vale.Def.Types_s.insert_nat64 (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0) (8\n `op_Multiply` va_get_reg64 rRsi va_s0) 1) (8 `op_Multiply` plain_num_bytes) 0) in let\n (raw_auth_quads:(seq quad32)) = va_if (va_get_reg64 rRsi va_s0 > va_get_reg64 rRdx va_s0\n `op_Multiply` 128 `op_Division` 8) (fun _ -> FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 7 va_s0) abytes_b)) (fun _ -> Vale.X64.Decls.s128 (va_get_mem_heaplet 1\n va_s0) auth_b) in let (auth_input_bytes:(seq nat8)) = FStar.Seq.Base.slice\n #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes raw_auth_quads) 0 (va_get_reg64\n rRsi va_s0) in let (padded_auth_bytes:(seq nat8)) = Vale.AES.GCTR_s.pad_to_128_bits\n auth_input_bytes in let (auth_quad_seq:(seq quad32)) = Vale.Def.Types_s.le_bytes_to_seq_quad32\n padded_auth_bytes in let (raw_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32\n (FStar.Seq.Base.append #quad32 auth_quad_seq (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0)\n in128x6_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b) in let (total_bytes:nat)\n = FStar.Seq.Base.length #quad32 auth_quad_seq `op_Multiply` 16 + plain_num_bytes in let\n (raw_quad_seq:(seq quad32)) = va_if (plain_num_bytes > (len128x6 + len128) `op_Multiply` 128\n `op_Division` 8) (fun _ -> let (ab:(seq nat8)) = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes (FStar.Seq.Base.append #quad32 raw_quad_seq\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b))) 0 total_bytes in let (pb:(seq\n nat8)) = Vale.AES.GCTR_s.pad_to_128_bits ab in Vale.Def.Types_s.le_bytes_to_seq_quad32 pb) (fun\n _ -> raw_quad_seq) in let (auth_quad_seq:(seq quad32)) = FStar.Seq.Base.append #quad32\n raw_quad_seq (FStar.Seq.Base.create #quad32 1 length_quad) in va_get_xmm 8 va_sM ==\n Vale.AES.GCTR_s.gctr_encrypt_block ctr_BE_1 (Vale.AES.GHash_s.ghash_LE h_LE auth_quad_seq) alg\n key 0))) ==> va_k va_sM (())))\n\n\nval va_wpProof_Gcm_blocks : alg:algorithm -> offset:int -> auth_b:buffer128 -> abytes_b:buffer128\n -> in128x6_b:buffer128 -> out128x6_b:buffer128 -> in128_b:buffer128 -> out128_b:buffer128 ->\n inout_b:buffer128 -> iv_b:buffer128 -> scratch_b:buffer128 -> key:(seq nat32) -> round_keys:(seq\n quad32) -> keys_b:buffer128 -> hkeys_b:buffer128 -> va_s0:va_state -> va_k:(va_state -> unit ->\n Type0)\n -> Ghost (va_state & va_fuel & unit)\n (requires (va_t_require va_s0 /\\ va_wp_Gcm_blocks alg offset auth_b abytes_b in128x6_b out128x6_b\n in128_b out128_b inout_b iv_b scratch_b key round_keys keys_b hkeys_b va_s0 va_k))\n (ensures (fun (va_sM, va_f0, va_g) -> va_t_ensure (va_code_Gcm_blocks alg offset) ([va_Mod_flags;\n va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet 3; va_Mod_mem_heaplet 2;\n va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13; va_Mod_xmm 12; va_Mod_xmm\n 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5;\n va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR15;\n va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64 rR11; va_Mod_reg64 rR10;\n va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64 rRsi; va_Mod_reg64 rRdi;\n va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64 rRax; va_Mod_mem]) va_s0\n va_k ((va_sM, va_f0, va_g))))\n[@\"opaque_to_smt\"]\nlet va_wpProof_Gcm_blocks alg offset auth_b abytes_b in128x6_b out128x6_b in128_b out128_b inout_b\n iv_b scratch_b key round_keys keys_b hkeys_b va_s0 va_k =\n let (va_sM, va_f0) = va_lemma_Gcm_blocks (va_code_Gcm_blocks alg offset) va_s0 alg offset auth_b\n abytes_b in128x6_b out128x6_b in128_b out128_b inout_b iv_b scratch_b key round_keys keys_b\n hkeys_b in\n va_lemma_upd_update va_sM;\n assert (va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 6 va_sM\n (va_update_mem_heaplet 5 va_sM (va_update_mem_heaplet 3 va_sM (va_update_mem_heaplet 2 va_sM\n (va_update_mem_heaplet 1 va_sM (va_update_xmm 15 va_sM (va_update_xmm 14 va_sM (va_update_xmm\n 13 va_sM (va_update_xmm 12 va_sM (va_update_xmm 11 va_sM (va_update_xmm 10 va_sM (va_update_xmm\n 9 va_sM (va_update_xmm 8 va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5\n va_sM (va_update_xmm 4 va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1\n va_sM (va_update_xmm 0 va_sM (va_update_reg64 rR15 va_sM (va_update_reg64 rR14 va_sM\n (va_update_reg64 rR13 va_sM (va_update_reg64 rR12 va_sM (va_update_reg64 rR11 va_sM\n (va_update_reg64 rR10 va_sM (va_update_reg64 rR9 va_sM (va_update_reg64 rR8 va_sM\n (va_update_reg64 rRbp va_sM (va_update_reg64 rRsi va_sM (va_update_reg64 rRdi va_sM\n (va_update_reg64 rRdx va_sM (va_update_reg64 rRcx va_sM (va_update_reg64 rRbx va_sM\n (va_update_reg64 rRax va_sM (va_update_ok va_sM (va_update_mem va_sM\n va_s0))))))))))))))))))))))))))))))))))))))));\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet\n 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13;\n va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm\n 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64\n rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64\n rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64\n rRax; va_Mod_mem]) va_sM va_s0;\n let va_g = () in\n (va_sM, va_f0, va_g)\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_quick_Gcm_blocks (alg:algorithm) (offset:int) (auth_b:buffer128) (abytes_b:buffer128)\n (in128x6_b:buffer128) (out128x6_b:buffer128) (in128_b:buffer128) (out128_b:buffer128)\n (inout_b:buffer128) (iv_b:buffer128) (scratch_b:buffer128) (key:(seq nat32)) (round_keys:(seq\n quad32)) (keys_b:buffer128) (hkeys_b:buffer128) : (va_quickCode unit (va_code_Gcm_blocks alg\n offset)) =\n (va_QProc (va_code_Gcm_blocks alg offset) ([va_Mod_flags; va_Mod_mem_heaplet 6;\n va_Mod_mem_heaplet 5; va_Mod_mem_heaplet 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1;\n va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13; va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10;\n va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm\n 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64\n rR13; va_Mod_reg64 rR12; va_Mod_reg64 rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64\n rR8; va_Mod_reg64 rRbp; va_Mod_reg64 rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64\n rRcx; va_Mod_reg64 rRbx; va_Mod_reg64 rRax; va_Mod_mem]) (va_wp_Gcm_blocks alg offset auth_b\n abytes_b in128x6_b out128x6_b in128_b out128_b inout_b iv_b scratch_b key round_keys keys_b\n hkeys_b) (va_wpProof_Gcm_blocks alg offset auth_b abytes_b in128x6_b out128x6_b in128_b\n out128_b inout_b iv_b scratch_b key round_keys keys_b hkeys_b))\n#pop-options\n//--\n//-- Gcm_blocks_wrapped\n#push-options \"--z3rlimit 60\"\n\nval va_code_Gcm_blocks_wrapped : alg:algorithm -> offset:int -> Tot va_code\n[@ \"opaque_to_smt\" va_qattr]\nlet va_code_Gcm_blocks_wrapped alg offset =\n (va_Block (va_CCons (va_code_Gcm_blocks alg offset) (va_CCons (va_Block (va_CNil ())) (va_CCons\n (va_Block (va_CNil ())) (va_CNil ())))))\n\nval va_codegen_success_Gcm_blocks_wrapped : alg:algorithm -> offset:int -> Tot va_pbool\n[@ \"opaque_to_smt\" va_qattr]\nlet va_codegen_success_Gcm_blocks_wrapped alg offset =\n (va_pbool_and (va_codegen_success_Gcm_blocks alg offset) (va_ttrue ()))\n\n[@ \"opaque_to_smt\" va_qattr]\nlet va_qcode_Gcm_blocks_wrapped (va_mods:va_mods_t) (alg:algorithm) (offset:int) (auth_b:buffer128)\n (abytes_b:buffer128) (in128x6_b:buffer128) (out128x6_b:buffer128) (in128_b:buffer128)\n (out128_b:buffer128) (inout_b:buffer128) (iv_b:buffer128) (iv:supported_iv_LE)\n (scratch_b:buffer128) (key:(seq nat32)) (round_keys:(seq quad32)) (keys_b:buffer128)\n (hkeys_b:buffer128) (expected_tag:(seq nat8)) : (va_quickCode unit (va_code_Gcm_blocks_wrapped\n alg offset)) =\n (qblock va_mods (fun (va_s:va_state) -> let (va_old_s:va_state) = va_s in let\n (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 0) (va_get_stack va_s) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 8) (va_get_stack va_s) in let\n (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 16) (va_get_stack va_s) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 24) (va_get_stack va_s) in let\n (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s +\n offset + 32) (va_get_stack va_s) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 40) (va_get_stack va_s) in let\n (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset\n + 48) (va_get_stack va_s) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s + offset + 56) (va_get_stack va_s) in let\n (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s\n + offset + 64) (va_get_stack va_s) in va_QBind va_range1\n \"***** PRECONDITION NOT MET AT line 739 column 15 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_quick_Gcm_blocks alg offset auth_b abytes_b in128x6_b out128x6_b in128_b out128_b inout_b\n iv_b scratch_b key round_keys keys_b hkeys_b) (fun (va_s:va_state) _ -> va_qAssertSquash\n va_range1\n \"***** EXPRESSION PRECONDITIONS NOT MET WITHIN line 741 column 37 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (8 `op_Multiply` va_get_reg64 rRsi va_old_s >= 0 /\\ 8 `op_Multiply` va_get_reg64 rRsi va_old_s\n <= 18446744073709551615 /\\ 8 `op_Multiply` plain_num_bytes >= 0 /\\ 8 `op_Multiply`\n plain_num_bytes <= 18446744073709551615) (fun _ -> let (va_arg55:Vale.Def.Types_s.quad32) =\n Vale.Def.Types_s.reverse_bytes_quad32 (Vale.Def.Types_s.insert_nat64\n (Vale.Def.Types_s.insert_nat64 (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0) (8\n `op_Multiply` va_get_reg64 rRsi va_old_s) 1) (8 `op_Multiply` plain_num_bytes) 0) in let\n (va_arg54:Vale.Def.Types_s.quad32) = va_get_xmm 8 va_s in let\n (va_arg53:Vale.Def.Types_s.quad32) = Vale.Def.Types_s.reverse_bytes_quad32\n (Vale.X64.Decls.buffer128_read hkeys_b 2 (va_get_mem_heaplet 0 va_s)) in let\n (va_arg52:Vale.Def.Types_s.quad32) = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2\n va_old_s) in let (va_arg51:Vale.AES.GCM_s.supported_iv_LE) = iv in let (va_arg50:Prims.nat) =\n va_get_reg64 rRsi va_old_s in let (va_arg49:Prims.nat) = plain_num_bytes in let\n (va_arg48:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s) inout_b in let (va_arg47:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) out128_b in let\n (va_arg46:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_s) out128x6_b in let (va_arg45:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_old_s) inout_b in let\n (va_arg44:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_old_s) in128_b in let (va_arg43:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b in\n let (va_arg42:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 7 va_old_s) abytes_b in let (va_arg41:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_old_s) auth_b in let\n (va_arg40:(FStar.Seq.Base.seq Vale.Def.Words_s.nat32)) = key in let\n (va_arg39:Vale.AES.AES_common_s.algorithm) = alg in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 741 column 37 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCM.gcm_blocks_dec_helper_simplified va_arg39 va_arg40 va_arg41\n va_arg42 va_arg43 va_arg44 va_arg45 va_arg46 va_arg47 va_arg48 va_arg49 va_arg50 va_arg51\n va_arg52 va_arg53 va_arg54 va_arg55) (let (auth_raw_quads:(FStar.Seq.Base.seq\n Vale.X64.Decls.quad32)) = FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_old_s) auth_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 7 va_old_s)\n abytes_b) in va_qAssertSquash va_range1\n \"***** EXPRESSION PRECONDITIONS NOT MET WITHIN line 751 column 5 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n ((fun a_1906 (s_1907:(FStar.Seq.Base.seq a_1906)) (i_1908:Prims.nat) (j_1909:Prims.nat) -> let\n (j_1869:Prims.nat) = j_1909 in Prims.b2t (Prims.op_AmpAmp (Prims.op_LessThanOrEqual i_1908\n j_1869) (Prims.op_LessThanOrEqual j_1869 (FStar.Seq.Base.length #a_1906 s_1907))))\n Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes auth_raw_quads) 0 (va_get_reg64\n rRsi va_old_s)) (fun _ -> let (auth_bytes:(FStar.Seq.Base.seq Vale.Def.Types_s.nat8)) =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n auth_raw_quads) 0 (va_get_reg64 rRsi va_old_s) in let (va_arg38:Vale.Def.Types_s.quad32) =\n Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_old_s) in let\n (va_arg37:Vale.AES.GCM_s.supported_iv_LE) = iv in let (va_arg36:Prims.nat) = plain_num_bytes in\n let (va_arg35:(FStar.Seq.Base.seq Vale.Def.Words_s.nat8)) = expected_tag in let\n (va_arg34:(FStar.Seq.Base.seq Vale.Def.Words_s.nat8)) = auth_bytes in let\n (va_arg33:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s) inout_b in let (va_arg32:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s) out128_b in let\n (va_arg31:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_s) out128x6_b in let (va_arg30:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_old_s) inout_b in let\n (va_arg29:(FStar.Seq.Base.seq Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_old_s) in128_b in let (va_arg28:(FStar.Seq.Base.seq\n Vale.Def.Types_s.quad32)) = Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_old_s) in128x6_b in\n let (va_arg27:(FStar.Seq.Base.seq Vale.Def.Words_s.nat32)) = key in let\n (va_arg26:Vale.AES.AES_common_s.algorithm) = alg in va_qPURE va_range1\n \"***** PRECONDITION NOT MET AT line 752 column 37 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (fun (_:unit) -> Vale.AES.GCM.gcm_blocks_helper_dec_simplified va_arg26 va_arg27 va_arg28\n va_arg29 va_arg30 va_arg31 va_arg32 va_arg33 va_arg34 va_arg35 va_arg36 va_arg37 va_arg38)\n (va_QEmpty (()))))))))\n\n\nval va_lemma_Gcm_blocks_wrapped : va_b0:va_code -> va_s0:va_state -> alg:algorithm -> offset:int ->\n auth_b:buffer128 -> abytes_b:buffer128 -> in128x6_b:buffer128 -> out128x6_b:buffer128 ->\n in128_b:buffer128 -> out128_b:buffer128 -> inout_b:buffer128 -> iv_b:buffer128 ->\n iv:supported_iv_LE -> scratch_b:buffer128 -> key:(seq nat32) -> round_keys:(seq quad32) ->\n keys_b:buffer128 -> hkeys_b:buffer128 -> expected_tag:(seq nat8)\n -> Ghost (va_state & va_fuel)\n (requires (va_require_total va_b0 (va_code_Gcm_blocks_wrapped alg offset) va_s0 /\\ va_get_ok\n va_s0 /\\ (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64\n rRsp va_s0 + offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in sse_enabled /\\ movbe_enabled /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0)\n Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp\n va_s0 + offset + 8) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 24) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 40) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64\n rRsp va_s0 + offset + 56) (va_get_stack va_s0) Public (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack\n va_s0) Public (va_get_stackTaint va_s0) /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet\n 1 va_s0) (va_get_reg64 rRdi va_s0) auth_b (va_get_reg64 rRdx va_s0) (va_get_mem_layout va_s0)\n Secret /\\ Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 7 va_s0) abytes_ptr abytes_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 2\n va_s0) (va_get_reg64 rR8 va_s0) iv_b 1 (va_get_mem_layout va_s0) Public /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 6 va_s0) in128x6_ptr in128x6_b len128x6\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 6\n va_s0) out128x6_ptr out128x6_b len128x6 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0) in128_ptr in128_b len128\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1\n va_s0) out128_ptr out128_b len128 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0) inout_ptr inout_b 1\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 3\n va_s0) (va_get_reg64 rRbp va_s0) scratch_b 9 (va_get_mem_layout va_s0) Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 0 va_s0) (va_get_reg64 rR9 va_s0) hkeys_b 8\n (va_get_mem_layout va_s0) Secret /\\ Vale.X64.Decls.buffer_disjoints128 iv_b ([keys_b;\n scratch_b; in128x6_b; out128x6_b; hkeys_b; in128_b; out128_b; inout_b]) /\\\n Vale.X64.Decls.buffer_disjoints128 scratch_b ([keys_b; in128x6_b; out128x6_b; in128_b;\n out128_b; inout_b; hkeys_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128x6_b ([keys_b;\n hkeys_b; in128_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 out128_b ([keys_b; hkeys_b;\n out128x6_b; inout_b]) /\\ Vale.X64.Decls.buffer_disjoints128 inout_b ([keys_b; hkeys_b;\n out128x6_b; out128_b]) /\\ (Vale.X64.Decls.buffers_disjoint128 in128x6_b out128x6_b \\/ in128x6_b\n == out128x6_b) /\\ (Vale.X64.Decls.buffers_disjoint128 in128_b out128_b \\/ in128_b == out128_b)\n /\\ va_get_reg64 rRdi va_s0 + 16 `op_Multiply` va_get_reg64 rRdx va_s0 < pow2_64 /\\ in128x6_ptr\n + 16 `op_Multiply` len128x6 < pow2_64 /\\ out128x6_ptr + 16 `op_Multiply` len128x6 < pow2_64 /\\\n in128_ptr + 16 `op_Multiply` len128 < pow2_64 /\\ out128_ptr + 16 `op_Multiply` len128 < pow2_64\n /\\ inout_ptr + 16 < pow2_64 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 auth_b ==\n va_get_reg64 rRdx va_s0 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 abytes_b == 1\n /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128x6_b ==\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out128x6_b /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b == Vale.X64.Decls.buffer_length\n #Vale.X64.Memory.vuint128 out128_b /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128\n in128x6_b == len128x6 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b ==\n len128 /\\ Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 inout_b == 1 /\\\n plain_num_bytes < pow2_32 /\\ va_get_reg64 rRsi va_s0 < pow2_32 /\\ va_get_reg64 rR9 va_s0 + 32 <\n pow2_64 /\\ Vale.X64.Memory.buffer_addr #Vale.X64.Memory.vuint128 keys_b (va_get_mem_heaplet 0\n va_s0) + 128 < pow2_64 /\\ len128x6 `op_Modulus` 6 == 0 /\\ (len128x6 > 0 ==> len128x6 >= 6) /\\\n 12 + len128x6 + 6 < pow2_32 /\\ (va_mul_nat len128x6 (128 `op_Division` 8) + va_mul_nat len128\n (128 `op_Division` 8) <= plain_num_bytes /\\ plain_num_bytes < va_mul_nat len128x6 (128\n `op_Division` 8) + va_mul_nat len128 (128 `op_Division` 8) + 128 `op_Division` 8) /\\\n (va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) <= va_get_reg64 rRsi va_s0 /\\\n va_get_reg64 rRsi va_s0 < va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) + 128\n `op_Division` 8) /\\ aes_reqs alg key round_keys keys_b (va_get_reg64 rRcx va_s0)\n (va_get_mem_heaplet 0 va_s0) (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 (Vale.AES.AES_s.aes_encrypt_LE alg key\n (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0))) /\\ (let iv_BE =\n Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in let h_LE =\n Vale.AES.AES_s.aes_encrypt_LE alg key (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0)\n in iv_BE == Vale.AES.GCM_s.compute_iv_BE h_LE iv))))\n (ensures (fun (va_sM, va_fM) -> va_ensure_total va_b0 va_s0 va_sM va_fM /\\ va_get_ok va_sM /\\\n (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 0) (va_get_stack va_s0) in let (in128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0) in\n let (out128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 16) (va_get_stack va_s0) in let (len128x6:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0) in\n let (in128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 32) (va_get_stack va_s0) in let (out128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0) in\n let (len128:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 48) (va_get_stack va_s0) in let (inout_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0) in\n let (plain_num_bytes:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp\n va_s0 + offset + 64) (va_get_stack va_s0) in Vale.X64.Decls.modifies_buffer128 out128_b\n (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1 va_sM) /\\ Vale.X64.Decls.modifies_buffer128\n iv_b (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2 va_sM) /\\\n Vale.X64.Decls.modifies_buffer128 scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3\n va_sM) /\\ Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0)\n (va_get_mem_heaplet 5 va_sM) /\\ Vale.X64.Decls.modifies_buffer128 out128x6_b\n (va_get_mem_heaplet 6 va_s0) (va_get_mem_heaplet 6 va_sM) /\\ plain_num_bytes < pow2_32 /\\\n va_get_reg64 rRsi va_s0 < pow2_32 /\\ (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0\n (va_get_mem_heaplet 2 va_s0) in let auth_raw_quads = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 7 va_s0) abytes_b) in let auth_bytes =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n auth_raw_quads) 0 (va_get_reg64 rRsi va_s0) in let plain_raw_quads = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_s0) in128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0)\n in128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b) in let plain_bytes =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n plain_raw_quads) 0 plain_num_bytes in let cipher_raw_quads = FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 6 va_sM) out128x6_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM)\n out128_b)) (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM) inout_b) in let cipher_bytes =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8 (Vale.Def.Types_s.le_seq_quad32_to_bytes\n cipher_raw_quads) 0 plain_num_bytes in l_and (l_and (l_and (FStar.Seq.Base.length\n #Vale.Def.Types_s.nat8 auth_bytes < pow2_32) (FStar.Seq.Base.length #Vale.Def.Types_s.nat8\n plain_bytes < pow2_32)) (cipher_bytes == __proj__Mktuple2__item___1 #(FStar.Seq.Base.seq\n Vale.Def.Types_s.nat8) #bool (Vale.AES.GCM_s.gcm_decrypt_LE alg\n (Vale.Def.Words.Seq_s.seq_nat32_to_seq_nat8_LE key) iv plain_bytes auth_bytes expected_tag)))\n (Vale.Def.Types_s.le_quad32_to_bytes (va_get_xmm 8 va_sM) == Vale.AES.GCM.gcm_decrypt_LE_tag\n alg (Vale.Def.Words.Seq_s.seq_nat32_to_seq_nat8_LE key) iv plain_bytes auth_bytes))) /\\\n va_state_eq va_sM (va_update_flags va_sM (va_update_mem_heaplet 6 va_sM (va_update_mem_heaplet\n 5 va_sM (va_update_mem_heaplet 3 va_sM (va_update_mem_heaplet 2 va_sM (va_update_mem_heaplet 1\n va_sM (va_update_xmm 15 va_sM (va_update_xmm 14 va_sM (va_update_xmm 13 va_sM (va_update_xmm 12\n va_sM (va_update_xmm 11 va_sM (va_update_xmm 10 va_sM (va_update_xmm 9 va_sM (va_update_xmm 8\n va_sM (va_update_xmm 7 va_sM (va_update_xmm 6 va_sM (va_update_xmm 5 va_sM (va_update_xmm 4\n va_sM (va_update_xmm 3 va_sM (va_update_xmm 2 va_sM (va_update_xmm 1 va_sM (va_update_xmm 0\n va_sM (va_update_reg64 rR15 va_sM (va_update_reg64 rR14 va_sM (va_update_reg64 rR13 va_sM\n (va_update_reg64 rR12 va_sM (va_update_reg64 rR11 va_sM (va_update_reg64 rR10 va_sM\n (va_update_reg64 rR9 va_sM (va_update_reg64 rR8 va_sM (va_update_reg64 rRbp va_sM\n (va_update_reg64 rRsi va_sM (va_update_reg64 rRdi va_sM (va_update_reg64 rRdx va_sM\n (va_update_reg64 rRcx va_sM (va_update_reg64 rRbx va_sM (va_update_reg64 rRax va_sM\n (va_update_ok va_sM (va_update_mem va_sM va_s0)))))))))))))))))))))))))))))))))))))))))\n[@\"opaque_to_smt\"]\nlet va_lemma_Gcm_blocks_wrapped va_b0 va_s0 alg offset auth_b abytes_b in128x6_b out128x6_b in128_b\n out128_b inout_b iv_b iv scratch_b key round_keys keys_b hkeys_b expected_tag =\n let (va_mods:va_mods_t) = [va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5;\n va_Mod_mem_heaplet 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14;\n va_Mod_xmm 13; va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8;\n va_Mod_xmm 7; va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm\n 1; va_Mod_xmm 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12;\n va_Mod_reg64 rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp;\n va_Mod_reg64 rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx;\n va_Mod_reg64 rRax; va_Mod_ok; va_Mod_mem] in\n let va_qc = va_qcode_Gcm_blocks_wrapped va_mods alg offset auth_b abytes_b in128x6_b out128x6_b\n in128_b out128_b inout_b iv_b iv scratch_b key round_keys keys_b hkeys_b expected_tag in\n let (va_sM, va_fM, va_g) = va_wp_sound_code_norm (va_code_Gcm_blocks_wrapped alg offset) va_qc\n va_s0 (fun va_s0 va_sM va_g -> let () = va_g in label va_range1\n \"***** POSTCONDITION NOT MET AT line 588 column 1 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_ok va_sM) /\\ (let (abytes_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64\n (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0) in let\n (in128x6_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 8) (va_get_stack va_s0) in let (out128x6_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack va_s0) in\n let (len128x6:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 +\n offset + 24) (va_get_stack va_s0) in let (in128_ptr:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack va_s0) in\n let (out128_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 40) (va_get_stack va_s0) in let (len128:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack va_s0) in\n let (inout_ptr:Vale.X64.Memory.nat64) = Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0\n + offset + 56) (va_get_stack va_s0) in let (plain_num_bytes:Vale.X64.Memory.nat64) =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack va_s0) in\n label va_range1\n \"***** POSTCONDITION NOT MET AT line 711 column 56 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out128_b (va_get_mem_heaplet 1 va_s0) (va_get_mem_heaplet 1\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 712 column 52 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 iv_b (va_get_mem_heaplet 2 va_s0) (va_get_mem_heaplet 2\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 713 column 57 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 scratch_b (va_get_mem_heaplet 3 va_s0) (va_get_mem_heaplet 3\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 714 column 55 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 inout_b (va_get_mem_heaplet 5 va_s0) (va_get_mem_heaplet 5\n va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 715 column 58 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (Vale.X64.Decls.modifies_buffer128 out128x6_b (va_get_mem_heaplet 6 va_s0) (va_get_mem_heaplet\n 6 va_sM)) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 718 column 39 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (plain_num_bytes < pow2_32) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 719 column 38 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (va_get_reg64 rRsi va_s0 < pow2_32) /\\ label va_range1\n \"***** POSTCONDITION NOT MET AT line 721 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 723 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let auth_raw_quads = FStar.Seq.Base.append #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128\n (va_get_mem_heaplet 1 va_s0) auth_b) (Vale.X64.Decls.s128 (va_get_mem_heaplet 7 va_s0)\n abytes_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 724 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let auth_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes auth_raw_quads) 0 (va_get_reg64 rRsi va_s0) in label\n va_range1\n \"***** POSTCONDITION NOT MET AT line 725 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let plain_raw_quads = FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_s0) inout_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 726 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let plain_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes plain_raw_quads) 0 plain_num_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 727 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let cipher_raw_quads = FStar.Seq.Base.append #Vale.X64.Decls.quad32 (FStar.Seq.Base.append\n #Vale.X64.Decls.quad32 (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM) out128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out128_b)) (Vale.X64.Decls.s128\n (va_get_mem_heaplet 5 va_sM) inout_b) in label va_range1\n \"***** POSTCONDITION NOT MET AT line 728 column 9 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (let cipher_bytes = FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes cipher_raw_quads) 0 plain_num_bytes in label va_range1\n \"***** POSTCONDITION NOT MET AT line 737 column 52 of file /home/gebner/fstar_dataset/projects/hacl-star/vale/code/crypto/aes/x64/Vale.AES.X64.GCMdecryptOpt.vaf *****\"\n (l_and (l_and (l_and (FStar.Seq.Base.length #Vale.Def.Types_s.nat8 auth_bytes < pow2_32)\n (FStar.Seq.Base.length #Vale.Def.Types_s.nat8 plain_bytes < pow2_32)) (cipher_bytes ==\n __proj__Mktuple2__item___1 #(FStar.Seq.Base.seq Vale.Def.Types_s.nat8) #bool\n (Vale.AES.GCM_s.gcm_decrypt_LE alg (Vale.Def.Words.Seq_s.seq_nat32_to_seq_nat8_LE key) iv\n plain_bytes auth_bytes expected_tag))) (Vale.Def.Types_s.le_quad32_to_bytes (va_get_xmm 8\n va_sM) == Vale.AES.GCM.gcm_decrypt_LE_tag alg (Vale.Def.Words.Seq_s.seq_nat32_to_seq_nat8_LE\n key) iv plain_bytes auth_bytes))))))))))) in\n assert_norm (va_qc.mods == va_mods);\n va_lemma_norm_mods ([va_Mod_flags; va_Mod_mem_heaplet 6; va_Mod_mem_heaplet 5; va_Mod_mem_heaplet\n 3; va_Mod_mem_heaplet 2; va_Mod_mem_heaplet 1; va_Mod_xmm 15; va_Mod_xmm 14; va_Mod_xmm 13;\n va_Mod_xmm 12; va_Mod_xmm 11; va_Mod_xmm 10; va_Mod_xmm 9; va_Mod_xmm 8; va_Mod_xmm 7;\n va_Mod_xmm 6; va_Mod_xmm 5; va_Mod_xmm 4; va_Mod_xmm 3; va_Mod_xmm 2; va_Mod_xmm 1; va_Mod_xmm\n 0; va_Mod_reg64 rR15; va_Mod_reg64 rR14; va_Mod_reg64 rR13; va_Mod_reg64 rR12; va_Mod_reg64\n rR11; va_Mod_reg64 rR10; va_Mod_reg64 rR9; va_Mod_reg64 rR8; va_Mod_reg64 rRbp; va_Mod_reg64\n rRsi; va_Mod_reg64 rRdi; va_Mod_reg64 rRdx; va_Mod_reg64 rRcx; va_Mod_reg64 rRbx; va_Mod_reg64\n rRax; va_Mod_ok; va_Mod_mem]) va_sM va_s0;\n (va_sM, va_fM)\n\nPartial implementation:\nlet va_wp_Gcm_blocks_wrapped\n (alg: algorithm)\n (offset: int)\n (auth_b abytes_b in128x6_b out128x6_b in128_b out128_b inout_b iv_b: buffer128)\n (iv: supported_iv_LE)\n (scratch_b: buffer128)\n (key: (seq nat32))\n (round_keys: (seq quad32))\n (keys_b hkeys_b: buffer128)\n (expected_tag: (seq nat8))\n (va_s0: va_state)\n (va_k: (va_state -> unit -> Type0))\n : Type0 =\n \n"
},
{
"role": "assistant",
"content": "(va_get_ok va_s0 /\\\n (let abytes_ptr:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0)\n in\n let in128x6_ptr:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0)\n in\n let out128x6_ptr:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack va_s0)\n in\n let len128x6:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0)\n in\n let in128_ptr:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack va_s0)\n in\n let out128_ptr:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0)\n in\n let len128:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack va_s0)\n in\n let inout_ptr:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0)\n in\n let plain_num_bytes:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack va_s0)\n in\n sse_enabled /\\ movbe_enabled /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 0)\n (va_get_stack va_s0)\n Public\n (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 8)\n (va_get_stack va_s0)\n Public\n (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 16)\n (va_get_stack va_s0)\n Public\n (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 24)\n (va_get_stack va_s0)\n Public\n (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 32)\n (va_get_stack va_s0)\n Public\n (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 40)\n (va_get_stack va_s0)\n Public\n (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 48)\n (va_get_stack va_s0)\n Public\n (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 56)\n (va_get_stack va_s0)\n Public\n (va_get_stackTaint va_s0) /\\\n Vale.X64.Stack_i.valid_stack_slot64 (va_get_reg64 rRsp va_s0 + offset + 64)\n (va_get_stack va_s0)\n Public\n (va_get_stackTaint va_s0) /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0)\n (va_get_reg64 rRdi va_s0)\n auth_b\n (va_get_reg64 rRdx va_s0)\n (va_get_mem_layout va_s0)\n Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 7 va_s0)\n abytes_ptr\n abytes_b\n 1\n (va_get_mem_layout va_s0)\n Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 2 va_s0)\n (va_get_reg64 rR8 va_s0)\n iv_b\n 1\n (va_get_mem_layout va_s0)\n Public /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 6 va_s0)\n in128x6_ptr\n in128x6_b\n len128x6\n (va_get_mem_layout va_s0)\n Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 6 va_s0)\n out128x6_ptr\n out128x6_b\n len128x6\n (va_get_mem_layout va_s0)\n Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 1 va_s0)\n in128_ptr\n in128_b\n len128\n (va_get_mem_layout va_s0)\n Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 1 va_s0)\n out128_ptr\n out128_b\n len128\n (va_get_mem_layout va_s0)\n Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 5 va_s0)\n inout_ptr\n inout_b\n 1\n (va_get_mem_layout va_s0)\n Secret /\\\n Vale.X64.Decls.validDstAddrs128 (va_get_mem_heaplet 3 va_s0)\n (va_get_reg64 rRbp va_s0)\n scratch_b\n 9\n (va_get_mem_layout va_s0)\n Secret /\\\n Vale.X64.Decls.validSrcAddrs128 (va_get_mem_heaplet 0 va_s0)\n (va_get_reg64 rR9 va_s0)\n hkeys_b\n 8\n (va_get_mem_layout va_s0)\n Secret /\\\n Vale.X64.Decls.buffer_disjoints128 iv_b\n ([keys_b; scratch_b; in128x6_b; out128x6_b; hkeys_b; in128_b; out128_b; inout_b]) /\\\n Vale.X64.Decls.buffer_disjoints128 scratch_b\n ([keys_b; in128x6_b; out128x6_b; in128_b; out128_b; inout_b; hkeys_b]) /\\\n Vale.X64.Decls.buffer_disjoints128 out128x6_b ([keys_b; hkeys_b; in128_b; inout_b]) /\\\n Vale.X64.Decls.buffer_disjoints128 out128_b ([keys_b; hkeys_b; out128x6_b; inout_b]) /\\\n Vale.X64.Decls.buffer_disjoints128 inout_b ([keys_b; hkeys_b; out128x6_b; out128_b]) /\\\n (Vale.X64.Decls.buffers_disjoint128 in128x6_b out128x6_b \\/ in128x6_b == out128x6_b) /\\\n (Vale.X64.Decls.buffers_disjoint128 in128_b out128_b \\/ in128_b == out128_b) /\\\n va_get_reg64 rRdi va_s0 + 16 `op_Multiply` (va_get_reg64 rRdx va_s0) < pow2_64 /\\\n in128x6_ptr + 16 `op_Multiply` len128x6 < pow2_64 /\\\n out128x6_ptr + 16 `op_Multiply` len128x6 < pow2_64 /\\\n in128_ptr + 16 `op_Multiply` len128 < pow2_64 /\\ out128_ptr + 16 `op_Multiply` len128 < pow2_64 /\\\n inout_ptr + 16 < pow2_64 /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 auth_b == va_get_reg64 rRdx va_s0 /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 abytes_b == 1 /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128x6_b ==\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out128x6_b /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b ==\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 out128_b /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128x6_b == len128x6 /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 in128_b == len128 /\\\n Vale.X64.Decls.buffer_length #Vale.X64.Memory.vuint128 inout_b == 1 /\\ plain_num_bytes < pow2_32 /\\\n va_get_reg64 rRsi va_s0 < pow2_32 /\\ va_get_reg64 rR9 va_s0 + 32 < pow2_64 /\\\n Vale.X64.Memory.buffer_addr #Vale.X64.Memory.vuint128 keys_b (va_get_mem_heaplet 0 va_s0) + 128 <\n pow2_64 /\\ len128x6 `op_Modulus` 6 == 0 /\\ (len128x6 > 0 ==> len128x6 >= 6) /\\\n 12 + len128x6 + 6 < pow2_32 /\\\n (va_mul_nat len128x6 (128 `op_Division` 8) + va_mul_nat len128 (128 `op_Division` 8) <=\n plain_num_bytes /\\\n plain_num_bytes <\n va_mul_nat len128x6 (128 `op_Division` 8) + va_mul_nat len128 (128 `op_Division` 8) +\n 128\n `op_Division`\n 8) /\\\n (va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) <= va_get_reg64 rRsi va_s0 /\\\n va_get_reg64 rRsi va_s0 <\n va_mul_nat (va_get_reg64 rRdx va_s0) (128 `op_Division` 8) + 128 `op_Division` 8) /\\\n aes_reqs alg\n key\n round_keys\n keys_b\n (va_get_reg64 rRcx va_s0)\n (va_get_mem_heaplet 0 va_s0)\n (va_get_mem_layout va_s0) /\\ pclmulqdq_enabled /\\\n Vale.AES.GHash.hkeys_reqs_priv (Vale.X64.Decls.s128 (va_get_mem_heaplet 0 va_s0) hkeys_b)\n (Vale.Def.Types_s.reverse_bytes_quad32 (Vale.AES.AES_s.aes_encrypt_LE alg\n key\n (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0))) /\\\n (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in\n let h_LE =\n Vale.AES.AES_s.aes_encrypt_LE alg\n key\n (Vale.Def.Words_s.Mkfour #Vale.Def.Types_s.nat32 0 0 0 0)\n in\n iv_BE == Vale.AES.GCM_s.compute_iv_BE h_LE iv)) /\\\n (forall (va_x_mem: vale_heap) (va_x_rax: nat64) (va_x_rbx: nat64) (va_x_rcx: nat64)\n (va_x_rdx: nat64) (va_x_rdi: nat64) (va_x_rsi: nat64) (va_x_rbp: nat64) (va_x_r8: nat64)\n (va_x_r9: nat64) (va_x_r10: nat64) (va_x_r11: nat64) (va_x_r12: nat64) (va_x_r13: nat64)\n (va_x_r14: nat64) (va_x_r15: nat64) (va_x_xmm0: quad32) (va_x_xmm1: quad32)\n (va_x_xmm2: quad32) (va_x_xmm3: quad32) (va_x_xmm4: quad32) (va_x_xmm5: quad32)\n (va_x_xmm6: quad32) (va_x_xmm7: quad32) (va_x_xmm8: quad32) (va_x_xmm9: quad32)\n (va_x_xmm10: quad32) (va_x_xmm11: quad32) (va_x_xmm12: quad32) (va_x_xmm13: quad32)\n (va_x_xmm14: quad32) (va_x_xmm15: quad32) (va_x_heap1: vale_heap) (va_x_heap2: vale_heap)\n (va_x_heap3: vale_heap) (va_x_heap5: vale_heap) (va_x_heap6: vale_heap)\n (va_x_efl: Vale.X64.Flags.t).\n let va_sM =\n va_upd_flags va_x_efl\n (va_upd_mem_heaplet 6\n va_x_heap6\n (va_upd_mem_heaplet 5\n va_x_heap5\n (va_upd_mem_heaplet 3\n va_x_heap3\n (va_upd_mem_heaplet 2\n va_x_heap2\n (va_upd_mem_heaplet 1\n va_x_heap1\n (va_upd_xmm 15\n va_x_xmm15\n (va_upd_xmm 14\n va_x_xmm14\n (va_upd_xmm 13\n va_x_xmm13\n (va_upd_xmm 12\n va_x_xmm12\n (va_upd_xmm 11\n va_x_xmm11\n (va_upd_xmm 10\n va_x_xmm10\n (va_upd_xmm 9\n va_x_xmm9\n (va_upd_xmm 8\n va_x_xmm8\n (va_upd_xmm 7\n va_x_xmm7\n (va_upd_xmm 6\n va_x_xmm6\n (va_upd_xmm 5\n va_x_xmm5\n (va_upd_xmm 4\n va_x_xmm4\n (va_upd_xmm 3\n va_x_xmm3\n (va_upd_xmm 2\n va_x_xmm2\n (va_upd_xmm 1\n va_x_xmm1\n (va_upd_xmm\n 0\n va_x_xmm0\n (va_upd_reg64\n rR15\n va_x_r15\n (va_upd_reg64\n rR14\n va_x_r14\n (\n va_upd_reg64\n rR13\n va_x_r13\n (\n va_upd_reg64\n rR12\n va_x_r12\n (\n va_upd_reg64\n rR11\n va_x_r11\n (\n va_upd_reg64\n rR10\n va_x_r10\n (\n va_upd_reg64\n rR9\n va_x_r9\n (\n va_upd_reg64\n rR8\n va_x_r8\n (\n va_upd_reg64\n rRbp\n va_x_rbp\n (\n va_upd_reg64\n rRsi\n va_x_rsi\n (\n va_upd_reg64\n rRdi\n va_x_rdi\n (\n va_upd_reg64\n rRdx\n va_x_rdx\n (\n va_upd_reg64\n rRcx\n va_x_rcx\n (\n va_upd_reg64\n rRbx\n va_x_rbx\n (\n va_upd_reg64\n rRax\n va_x_rax\n (\n va_upd_mem\n va_x_mem\n va_s0\n\n )\n\n )\n\n )\n\n )\n\n )\n\n )\n\n )\n\n )\n\n )\n\n )\n\n )\n\n )\n\n )\n\n )\n\n ))\n ))))))))))\n )))))))))))\n in\n va_get_ok va_sM /\\\n (let abytes_ptr:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 0) (va_get_stack va_s0)\n in\n let in128x6_ptr:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 8) (va_get_stack va_s0)\n in\n let out128x6_ptr:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 16) (va_get_stack va_s0)\n in\n let len128x6:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 24) (va_get_stack va_s0)\n in\n let in128_ptr:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 32) (va_get_stack va_s0)\n in\n let out128_ptr:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 40) (va_get_stack va_s0)\n in\n let len128:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 48) (va_get_stack va_s0)\n in\n let inout_ptr:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 56) (va_get_stack va_s0)\n in\n let plain_num_bytes:Vale.X64.Memory.nat64 =\n Vale.X64.Stack_i.load_stack64 (va_get_reg64 rRsp va_s0 + offset + 64) (va_get_stack va_s0)\n in\n Vale.X64.Decls.modifies_buffer128 out128_b\n (va_get_mem_heaplet 1 va_s0)\n (va_get_mem_heaplet 1 va_sM) /\\\n Vale.X64.Decls.modifies_buffer128 iv_b\n (va_get_mem_heaplet 2 va_s0)\n (va_get_mem_heaplet 2 va_sM) /\\\n Vale.X64.Decls.modifies_buffer128 scratch_b\n (va_get_mem_heaplet 3 va_s0)\n (va_get_mem_heaplet 3 va_sM) /\\\n Vale.X64.Decls.modifies_buffer128 inout_b\n (va_get_mem_heaplet 5 va_s0)\n (va_get_mem_heaplet 5 va_sM) /\\\n Vale.X64.Decls.modifies_buffer128 out128x6_b\n (va_get_mem_heaplet 6 va_s0)\n (va_get_mem_heaplet 6 va_sM) /\\ plain_num_bytes < pow2_32 /\\\n va_get_reg64 rRsi va_s0 < pow2_32 /\\\n (let iv_BE = Vale.X64.Decls.buffer128_read iv_b 0 (va_get_mem_heaplet 2 va_s0) in\n let auth_raw_quads =\n FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) auth_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 7 va_s0) abytes_b)\n in\n let auth_bytes =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes auth_raw_quads)\n 0\n (va_get_reg64 rRsi va_s0)\n in\n let plain_raw_quads =\n FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_s0) in128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_s0) in128_b))\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_s0) inout_b)\n in\n let plain_bytes =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes plain_raw_quads)\n 0\n plain_num_bytes\n in\n let cipher_raw_quads =\n FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (FStar.Seq.Base.append #Vale.X64.Decls.quad32\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 6 va_sM) out128x6_b)\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 1 va_sM) out128_b))\n (Vale.X64.Decls.s128 (va_get_mem_heaplet 5 va_sM) inout_b)\n in\n let cipher_bytes =\n FStar.Seq.Base.slice #Vale.Def.Types_s.nat8\n (Vale.Def.Types_s.le_seq_quad32_to_bytes cipher_raw_quads)\n 0\n plain_num_bytes\n in\n l_and (l_and (l_and (FStar.Seq.Base.length #Vale.Def.Types_s.nat8 auth_bytes < pow2_32)\n (FStar.Seq.Base.length #Vale.Def.Types_s.nat8 plain_bytes < pow2_32))\n (cipher_bytes ==\n __proj__Mktuple2__item___1 #(FStar.Seq.Base.seq Vale.Def.Types_s.nat8)\n #bool\n (Vale.AES.GCM_s.gcm_decrypt_LE alg\n (Vale.Def.Words.Seq_s.seq_nat32_to_seq_nat8_LE key)\n iv\n plain_bytes\n auth_bytes\n expected_tag)))\n (Vale.Def.Types_s.le_quad32_to_bytes (va_get_xmm 8 va_sM) ==\n Vale.AES.GCM.gcm_decrypt_LE_tag alg\n (Vale.Def.Words.Seq_s.seq_nat32_to_seq_nat8_LE key)\n iv\n plain_bytes\n auth_bytes))) ==>\n va_k va_sM (())))"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: l1: AlgHeap.ops -> l2: AlgHeap.ops\n -> FStar.Pervasives.Lemma (ensures AlgHeap.sublist l1 (l1 @ l2) /\\ AlgHeap.sublist l2 (l1 @ l2))\n [SMTPatOr [[SMTPat (AlgHeap.sublist l1 (l1 @ l2))]; [SMTPat (AlgHeap.sublist l2 (l1 @ l2))]]]\nEffect: FStar.Pervasives.Lemma\n\nContext:\nmodule AlgHeap\n\n(* Essentially a copy of AlgForAll but using a heap for the state *)\n\nopen FStar.Tactics.V2\nopen FStar.List.Tot\nopen FStar.Universe\nmodule L = Lattice\nmodule Ghost = FStar.Ghost\nmodule Map = FStar.Map\nmodule T = FStar.Tactics.V2\n\ntype loc = int\ntype state = Map.t loc int\n\ntype empty : Type u#aa =\n\ntype op =\n | Read\n | Write\n | Raise\n | Other of int\n\nassume val other_inp : int -> Type u#0\nlet op_inp : op -> Type u#0 =\n function\n | Read -> unit\n | Write -> state\n | Raise -> exn\n | Other i -> other_inp i\n\nassume val other_out : int -> Type u#0\nlet op_out : op -> Type =\n function\n | Read -> state\n | Write -> unit\n | Raise -> empty\n | Other i -> other_inp i\n\nnoeq\ntype tree0 (a:Type) : Type =\n | Return : a -> tree0 a\n | Op : op:op -> i:(op_inp op) -> k:(op_out op -> tree0 a) -> tree0 a\n\ntype ops = list op\n\nlet rec abides #a (labs:ops) (f : tree0 a) : prop =\n begin match f with\n | Op a i k ->\n mem a labs /\\ (forall o. abides labs (k o))\n | Return _ -> True\n end\n\ntype tree (a:Type u#aa)\n (labs : list u#0 op)\n : Type u#aa\n =\n r:(tree0 a){abides labs r}\n\nlet rec interp_at (l1 l2 : ops) (l : op)\n : Lemma (mem l (l1@l2) == (mem l l1 || mem l l2))\n [SMTPat (mem l (l1@l2))]\n = match l1 with\n | [] -> ()\n | _::l1 -> interp_at l1 l2 l\n\nlet sublist (l1 l2 : ops) = forall x. mem x l1 ==> mem x l2\n\nPartial implementation:\nlet rec sublist_at (l1 l2: ops)\n : Lemma (sublist l1 (l1 @ l2) /\\ sublist l2 (l1 @ l2))\n [SMTPatOr [[SMTPat (sublist l1 (l1 @ l2))]; [SMTPat (sublist l2 (l1 @ l2))]]] =\n \n"
},
{
"role": "assistant",
"content": "match l1 with\n| [] -> ()\n| _ :: l1 -> sublist_at l1 l2"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: l1: AlgHeap.ops -> l2: AlgHeap.ops -> c: AlgHeap.tree0 a\n -> FStar.Pervasives.Lemma (requires AlgHeap.abides l1 c \\/ AlgHeap.abides l2 c)\n (ensures AlgHeap.abides (l1 @ l2) c)\n [SMTPat (AlgHeap.abides (l1 @ l2) c)]\nEffect: FStar.Pervasives.Lemma\n\nContext:\nmodule AlgHeap\n\n(* Essentially a copy of AlgForAll but using a heap for the state *)\n\nopen FStar.Tactics.V2\nopen FStar.List.Tot\nopen FStar.Universe\nmodule L = Lattice\nmodule Ghost = FStar.Ghost\nmodule Map = FStar.Map\nmodule T = FStar.Tactics.V2\n\ntype loc = int\ntype state = Map.t loc int\n\ntype empty : Type u#aa =\n\ntype op =\n | Read\n | Write\n | Raise\n | Other of int\n\nassume val other_inp : int -> Type u#0\nlet op_inp : op -> Type u#0 =\n function\n | Read -> unit\n | Write -> state\n | Raise -> exn\n | Other i -> other_inp i\n\nassume val other_out : int -> Type u#0\nlet op_out : op -> Type =\n function\n | Read -> state\n | Write -> unit\n | Raise -> empty\n | Other i -> other_inp i\n\nnoeq\ntype tree0 (a:Type) : Type =\n | Return : a -> tree0 a\n | Op : op:op -> i:(op_inp op) -> k:(op_out op -> tree0 a) -> tree0 a\n\ntype ops = list op\n\nlet rec abides #a (labs:ops) (f : tree0 a) : prop =\n begin match f with\n | Op a i k ->\n mem a labs /\\ (forall o. abides labs (k o))\n | Return _ -> True\n end\n\ntype tree (a:Type u#aa)\n (labs : list u#0 op)\n : Type u#aa\n =\n r:(tree0 a){abides labs r}\n\nlet rec interp_at (l1 l2 : ops) (l : op)\n : Lemma (mem l (l1@l2) == (mem l l1 || mem l l2))\n [SMTPat (mem l (l1@l2))]\n = match l1 with\n | [] -> ()\n | _::l1 -> interp_at l1 l2 l\n\nlet sublist (l1 l2 : ops) = forall x. mem x l1 ==> mem x l2\n\nlet rec sublist_at\n (l1 l2 : ops)\n : Lemma (sublist l1 (l1@l2) /\\ sublist l2 (l1@l2))\n [SMTPatOr [[SMTPat (sublist l1 (l1@l2))];\n [SMTPat (sublist l2 (l1@l2))]]]\n = match l1 with\n | [] -> ()\n | _::l1 -> sublist_at l1 l2\n\nlet sublist_at_self\n (l : ops)\n : Lemma (sublist l (l@l))\n [SMTPat (sublist l (l@l))]\n = ()\n \nlet rec abides_sublist_nopat #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c) /\\ sublist l1 l2)\n (ensures (abides l2) c)\n = match c with\n | Return _ -> ()\n | Op a i k ->\n let sub o : Lemma (abides l2 (k o)) =\n abides_sublist_nopat l1 l2 (k o)\n in\n Classical.forall_intro sub\n\nlet abides_sublist #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c) /\\ sublist l1 l2)\n (ensures (abides l2 c))\n [SMTPat (abides l2 c); SMTPat (sublist l1 l2)]\n = abides_sublist_nopat l1 l2 c\n \nlet abides_at_self #a\n (l : ops)\n (c : tree0 a)\n : Lemma (abides (l@l) c <==> abides l c)\n [SMTPat (abides (l@l) c)]\n = (* Trigger some patterns *)\n assert (sublist l (l@l));\n assert (sublist (l@l) l)\n\nPartial implementation:\nlet abides_app #a (l1: ops) (l2: ops) (c: tree0 a)\n : Lemma (requires (abides l1 c \\/ abides l2 c))\n (ensures (abides (l1 @ l2) c))\n [SMTPat (abides (l1 @ l2) c)] =\n \n"
},
{
"role": "assistant",
"content": "sublist_at l1 l2"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: l1: AlgHeap.ops -> l2: AlgHeap.ops -> l: AlgHeap.op\n -> FStar.Pervasives.Lemma\n (ensures\n FStar.List.Tot.Base.mem l (l1 @ l2) ==\n (FStar.List.Tot.Base.mem l l1 || FStar.List.Tot.Base.mem l l2))\n [SMTPat (FStar.List.Tot.Base.mem l (l1 @ l2))]\nEffect: FStar.Pervasives.Lemma\n\nContext:\nmodule AlgHeap\n\n(* Essentially a copy of AlgForAll but using a heap for the state *)\n\nopen FStar.Tactics.V2\nopen FStar.List.Tot\nopen FStar.Universe\nmodule L = Lattice\nmodule Ghost = FStar.Ghost\nmodule Map = FStar.Map\nmodule T = FStar.Tactics.V2\n\ntype loc = int\ntype state = Map.t loc int\n\ntype empty : Type u#aa =\n\ntype op =\n | Read\n | Write\n | Raise\n | Other of int\n\nassume val other_inp : int -> Type u#0\nlet op_inp : op -> Type u#0 =\n function\n | Read -> unit\n | Write -> state\n | Raise -> exn\n | Other i -> other_inp i\n\nassume val other_out : int -> Type u#0\nlet op_out : op -> Type =\n function\n | Read -> state\n | Write -> unit\n | Raise -> empty\n | Other i -> other_inp i\n\nnoeq\ntype tree0 (a:Type) : Type =\n | Return : a -> tree0 a\n | Op : op:op -> i:(op_inp op) -> k:(op_out op -> tree0 a) -> tree0 a\n\ntype ops = list op\n\nlet rec abides #a (labs:ops) (f : tree0 a) : prop =\n begin match f with\n | Op a i k ->\n mem a labs /\\ (forall o. abides labs (k o))\n | Return _ -> True\n end\n\ntype tree (a:Type u#aa)\n (labs : list u#0 op)\n : Type u#aa\n =\n r:(tree0 a){abides labs r}\n\nPartial implementation:\nlet rec interp_at (l1 l2: ops) (l: op)\n : Lemma (mem l (l1 @ l2) == (mem l l1 || mem l l2)) [SMTPat (mem l (l1 @ l2))] =\n \n"
},
{
"role": "assistant",
"content": "match l1 with\n| [] -> ()\n| _ :: l1 -> interp_at l1 l2 l"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: t: Hacl.Bignum.Definitions.limb_t -> len: Hacl.Bignum.meta_len t -> Type0\nEffect: Prims.Tot\n\nContext:\nmodule Hacl.Bignum.ModInv\n\nopen FStar.HyperStack\nopen FStar.HyperStack.ST\nopen FStar.Mul\n\nopen Lib.IntTypes\nopen Lib.Buffer\n\nopen Hacl.Bignum.Definitions\n\nmodule ST = FStar.HyperStack.ST\nmodule S = Hacl.Spec.Bignum.ModInv\n\nmodule BN = Hacl.Bignum\nmodule BE = Hacl.Bignum.Exponentiation\nmodule BM = Hacl.Bignum.Montgomery\nmodule SD = Hacl.Spec.Bignum.Definitions\n\n#reset-options \"--z3rlimit 50 --fuel 0 --ifuel 0\"\n\ninline_for_extraction noextract\nlet bn_check_mod_inv_prime_st (t:limb_t) (len:BN.meta_len t) =\n n:lbignum t len\n -> a:lbignum t len ->\n Stack (limb t)\n (requires fun h -> live h n /\\ live h a /\\ disjoint n a)\n (ensures fun h0 r h1 -> modifies0 h0 h1 /\\\n r == S.bn_check_mod_inv_prime (as_seq h0 n) (as_seq h0 a))\n\n\ninline_for_extraction noextract\nval bn_check_mod_inv_prime: #t:limb_t -> len:BN.meta_len t -> bn_check_mod_inv_prime_st t len\nlet bn_check_mod_inv_prime #t len n a =\n let m0 = BM.bn_check_modulus n in\n let m1 = BN.bn_is_zero_mask len a in\n let m2 = BN.bn_lt_mask len a n in\n m0 &. (lognot m1) &. m2\n\n\ninline_for_extraction noextract\nval bn_mod_inv_prime_n2:\n #t:limb_t\n -> len:BN.meta_len t\n -> n:lbignum t len\n -> res:lbignum t len ->\n Stack unit\n (requires fun h -> 1 < bn_v h n /\\\n live h n /\\ live h res /\\ disjoint res n)\n (ensures fun h0 _ h1 -> modifies (loc res) h0 h1 /\\\n as_seq h1 res == S.bn_mod_inv_prime_n2 (as_seq h0 n))\n\nlet bn_mod_inv_prime_n2 #t len n res =\n let c = BN.bn_sub1 len n (uint #t #SEC 2) res in\n LowStar.Ignore.ignore c;\n ()\n\nPartial implementation:\nlet bn_mod_inv_prime_precomp_st (t: limb_t) (len: BN.meta_len t) =\n \n"
},
{
"role": "assistant",
"content": "n: lbignum t len -> mu: limb t -> r2: lbignum t len -> a: lbignum t len -> res: lbignum t len\n -> Stack unit\n (requires\n fun h ->\n live h n /\\ live h r2 /\\ live h a /\\ live h res /\\ disjoint res n /\\ disjoint res a /\\\n disjoint n a /\\ disjoint res r2 /\\ disjoint a r2 /\\ disjoint n r2 /\\\n S.bn_mod_inv_prime_pre (as_seq h n) (as_seq h a) /\\\n bn_v h r2 == pow2 ((2 * bits t) * v len) % bn_v h n /\\\n (1 + bn_v h n * v mu) % pow2 (bits t) == 0)\n (ensures fun h0 r h1 -> modifies (loc res) h0 h1 /\\ bn_v h1 res * bn_v h0 a % bn_v h0 n = 1)"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: l: AlgHeap.ops -> c: AlgHeap.tree0 a\n -> FStar.Pervasives.Lemma (ensures AlgHeap.abides (l @ l) c <==> AlgHeap.abides l c)\n [SMTPat (AlgHeap.abides (l @ l) c)]\nEffect: FStar.Pervasives.Lemma\n\nContext:\nmodule AlgHeap\n\n(* Essentially a copy of AlgForAll but using a heap for the state *)\n\nopen FStar.Tactics.V2\nopen FStar.List.Tot\nopen FStar.Universe\nmodule L = Lattice\nmodule Ghost = FStar.Ghost\nmodule Map = FStar.Map\nmodule T = FStar.Tactics.V2\n\ntype loc = int\ntype state = Map.t loc int\n\ntype empty : Type u#aa =\n\ntype op =\n | Read\n | Write\n | Raise\n | Other of int\n\nassume val other_inp : int -> Type u#0\nlet op_inp : op -> Type u#0 =\n function\n | Read -> unit\n | Write -> state\n | Raise -> exn\n | Other i -> other_inp i\n\nassume val other_out : int -> Type u#0\nlet op_out : op -> Type =\n function\n | Read -> state\n | Write -> unit\n | Raise -> empty\n | Other i -> other_inp i\n\nnoeq\ntype tree0 (a:Type) : Type =\n | Return : a -> tree0 a\n | Op : op:op -> i:(op_inp op) -> k:(op_out op -> tree0 a) -> tree0 a\n\ntype ops = list op\n\nlet rec abides #a (labs:ops) (f : tree0 a) : prop =\n begin match f with\n | Op a i k ->\n mem a labs /\\ (forall o. abides labs (k o))\n | Return _ -> True\n end\n\ntype tree (a:Type u#aa)\n (labs : list u#0 op)\n : Type u#aa\n =\n r:(tree0 a){abides labs r}\n\nlet rec interp_at (l1 l2 : ops) (l : op)\n : Lemma (mem l (l1@l2) == (mem l l1 || mem l l2))\n [SMTPat (mem l (l1@l2))]\n = match l1 with\n | [] -> ()\n | _::l1 -> interp_at l1 l2 l\n\nlet sublist (l1 l2 : ops) = forall x. mem x l1 ==> mem x l2\n\nlet rec sublist_at\n (l1 l2 : ops)\n : Lemma (sublist l1 (l1@l2) /\\ sublist l2 (l1@l2))\n [SMTPatOr [[SMTPat (sublist l1 (l1@l2))];\n [SMTPat (sublist l2 (l1@l2))]]]\n = match l1 with\n | [] -> ()\n | _::l1 -> sublist_at l1 l2\n\nlet sublist_at_self\n (l : ops)\n : Lemma (sublist l (l@l))\n [SMTPat (sublist l (l@l))]\n = ()\n \nlet rec abides_sublist_nopat #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c) /\\ sublist l1 l2)\n (ensures (abides l2) c)\n = match c with\n | Return _ -> ()\n | Op a i k ->\n let sub o : Lemma (abides l2 (k o)) =\n abides_sublist_nopat l1 l2 (k o)\n in\n Classical.forall_intro sub\n\nlet abides_sublist #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c) /\\ sublist l1 l2)\n (ensures (abides l2 c))\n [SMTPat (abides l2 c); SMTPat (sublist l1 l2)]\n = abides_sublist_nopat l1 l2 c\n\nPartial implementation:\nlet abides_at_self #a (l: ops) (c: tree0 a)\n : Lemma (abides (l @ l) c <==> abides l c) [SMTPat (abides (l @ l) c)] =\n \n"
},
{
"role": "assistant",
"content": "assert (sublist l (l @ l));\nassert (sublist (l @ l) l)"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: a: Type -> x: a -> AlgHeap.tree a []\nEffect: Prims.Tot\n\nContext:\nmodule AlgHeap\n\n(* Essentially a copy of AlgForAll but using a heap for the state *)\n\nopen FStar.Tactics.V2\nopen FStar.List.Tot\nopen FStar.Universe\nmodule L = Lattice\nmodule Ghost = FStar.Ghost\nmodule Map = FStar.Map\nmodule T = FStar.Tactics.V2\n\ntype loc = int\ntype state = Map.t loc int\n\ntype empty : Type u#aa =\n\ntype op =\n | Read\n | Write\n | Raise\n | Other of int\n\nassume val other_inp : int -> Type u#0\nlet op_inp : op -> Type u#0 =\n function\n | Read -> unit\n | Write -> state\n | Raise -> exn\n | Other i -> other_inp i\n\nassume val other_out : int -> Type u#0\nlet op_out : op -> Type =\n function\n | Read -> state\n | Write -> unit\n | Raise -> empty\n | Other i -> other_inp i\n\nnoeq\ntype tree0 (a:Type) : Type =\n | Return : a -> tree0 a\n | Op : op:op -> i:(op_inp op) -> k:(op_out op -> tree0 a) -> tree0 a\n\ntype ops = list op\n\nlet rec abides #a (labs:ops) (f : tree0 a) : prop =\n begin match f with\n | Op a i k ->\n mem a labs /\\ (forall o. abides labs (k o))\n | Return _ -> True\n end\n\ntype tree (a:Type u#aa)\n (labs : list u#0 op)\n : Type u#aa\n =\n r:(tree0 a){abides labs r}\n\nlet rec interp_at (l1 l2 : ops) (l : op)\n : Lemma (mem l (l1@l2) == (mem l l1 || mem l l2))\n [SMTPat (mem l (l1@l2))]\n = match l1 with\n | [] -> ()\n | _::l1 -> interp_at l1 l2 l\n\nlet sublist (l1 l2 : ops) = forall x. mem x l1 ==> mem x l2\n\nlet rec sublist_at\n (l1 l2 : ops)\n : Lemma (sublist l1 (l1@l2) /\\ sublist l2 (l1@l2))\n [SMTPatOr [[SMTPat (sublist l1 (l1@l2))];\n [SMTPat (sublist l2 (l1@l2))]]]\n = match l1 with\n | [] -> ()\n | _::l1 -> sublist_at l1 l2\n\nlet sublist_at_self\n (l : ops)\n : Lemma (sublist l (l@l))\n [SMTPat (sublist l (l@l))]\n = ()\n \nlet rec abides_sublist_nopat #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c) /\\ sublist l1 l2)\n (ensures (abides l2) c)\n = match c with\n | Return _ -> ()\n | Op a i k ->\n let sub o : Lemma (abides l2 (k o)) =\n abides_sublist_nopat l1 l2 (k o)\n in\n Classical.forall_intro sub\n\nlet abides_sublist #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c) /\\ sublist l1 l2)\n (ensures (abides l2 c))\n [SMTPat (abides l2 c); SMTPat (sublist l1 l2)]\n = abides_sublist_nopat l1 l2 c\n \nlet abides_at_self #a\n (l : ops)\n (c : tree0 a)\n : Lemma (abides (l@l) c <==> abides l c)\n [SMTPat (abides (l@l) c)]\n = (* Trigger some patterns *)\n assert (sublist l (l@l));\n assert (sublist (l@l) l)\n\nlet abides_app #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c \\/ abides l2 c))\n (ensures (abides (l1@l2) c))\n [SMTPat (abides (l1@l2) c)]\n = sublist_at l1 l2\n\n(* Folding a computation tree *)\nval fold_with (#a #b:_) (#labs : ops)\n (f:tree a labs)\n (v : a -> b)\n (h: (o:op{mem o labs} -> op_inp o -> (op_out o -> b) -> b))\n : b\nlet rec fold_with #a #b #labs f v h =\n match f with\n | Return x -> v x\n | Op act i k ->\n let k' (o : op_out act) : b =\n fold_with #_ #_ #labs (k o) v h\n in\n h act i k'\n\nlet handler_ty_l (o:op) (b:Type) (labs:ops) =\n op_inp o -> (op_out o -> tree b labs) -> tree b labs\n\nlet handler_ty (labs0 : ops) (b:Type) (labs1 : ops) : Type =\n o:op{mem o labs0} -> handler_ty_l o b labs1\n\n(* The most generic handling construct, we use it to implement bind.\nIt is actually just a special case of folding. *)\nval handle_with (#a #b:_) (#labs0 #labs1 : ops)\n (f:tree a labs0)\n (v : a -> tree b labs1)\n (h: handler_ty labs0 b labs1)\n : tree b labs1\nlet handle_with f v h = fold_with f v h\n\nPartial implementation:\nlet return (a: Type) (x: a) : tree a [] =\n \n"
},
{
"role": "assistant",
"content": "Return x"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: l1: AlgHeap.ops -> l2: AlgHeap.ops -> c: AlgHeap.tree0 a\n -> FStar.Pervasives.Lemma (requires AlgHeap.abides l1 c /\\ AlgHeap.sublist l1 l2)\n (ensures AlgHeap.abides l2 c)\nEffect: FStar.Pervasives.Lemma\n\nContext:\nmodule AlgHeap\n\n(* Essentially a copy of AlgForAll but using a heap for the state *)\n\nopen FStar.Tactics.V2\nopen FStar.List.Tot\nopen FStar.Universe\nmodule L = Lattice\nmodule Ghost = FStar.Ghost\nmodule Map = FStar.Map\nmodule T = FStar.Tactics.V2\n\ntype loc = int\ntype state = Map.t loc int\n\ntype empty : Type u#aa =\n\ntype op =\n | Read\n | Write\n | Raise\n | Other of int\n\nassume val other_inp : int -> Type u#0\nlet op_inp : op -> Type u#0 =\n function\n | Read -> unit\n | Write -> state\n | Raise -> exn\n | Other i -> other_inp i\n\nassume val other_out : int -> Type u#0\nlet op_out : op -> Type =\n function\n | Read -> state\n | Write -> unit\n | Raise -> empty\n | Other i -> other_inp i\n\nnoeq\ntype tree0 (a:Type) : Type =\n | Return : a -> tree0 a\n | Op : op:op -> i:(op_inp op) -> k:(op_out op -> tree0 a) -> tree0 a\n\ntype ops = list op\n\nlet rec abides #a (labs:ops) (f : tree0 a) : prop =\n begin match f with\n | Op a i k ->\n mem a labs /\\ (forall o. abides labs (k o))\n | Return _ -> True\n end\n\ntype tree (a:Type u#aa)\n (labs : list u#0 op)\n : Type u#aa\n =\n r:(tree0 a){abides labs r}\n\nlet rec interp_at (l1 l2 : ops) (l : op)\n : Lemma (mem l (l1@l2) == (mem l l1 || mem l l2))\n [SMTPat (mem l (l1@l2))]\n = match l1 with\n | [] -> ()\n | _::l1 -> interp_at l1 l2 l\n\nlet sublist (l1 l2 : ops) = forall x. mem x l1 ==> mem x l2\n\nlet rec sublist_at\n (l1 l2 : ops)\n : Lemma (sublist l1 (l1@l2) /\\ sublist l2 (l1@l2))\n [SMTPatOr [[SMTPat (sublist l1 (l1@l2))];\n [SMTPat (sublist l2 (l1@l2))]]]\n = match l1 with\n | [] -> ()\n | _::l1 -> sublist_at l1 l2\n\nlet sublist_at_self\n (l : ops)\n : Lemma (sublist l (l@l))\n [SMTPat (sublist l (l@l))]\n = ()\n\nPartial implementation:\nlet rec abides_sublist_nopat #a (l1: ops) (l2: ops) (c: tree0 a)\n : Lemma (requires (abides l1 c) /\\ sublist l1 l2) (ensures (abides l2) c) =\n \n"
},
{
"role": "assistant",
"content": "match c with\n| Return _ -> ()\n| Op a i k ->\n let sub o : Lemma (abides l2 (k o)) = abides_sublist_nopat l1 l2 (k o) in\n Classical.forall_intro sub"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: \n len: Hacl.Bignum.meta_len t ->\n n: Hacl.Bignum.Definitions.lbignum t len ->\n res: Hacl.Bignum.Definitions.lbignum t len\n -> FStar.HyperStack.ST.Stack Prims.unit\nEffect: FStar.HyperStack.ST.Stack\n\nContext:\nmodule Hacl.Bignum.ModInv\n\nopen FStar.HyperStack\nopen FStar.HyperStack.ST\nopen FStar.Mul\n\nopen Lib.IntTypes\nopen Lib.Buffer\n\nopen Hacl.Bignum.Definitions\n\nmodule ST = FStar.HyperStack.ST\nmodule S = Hacl.Spec.Bignum.ModInv\n\nmodule BN = Hacl.Bignum\nmodule BE = Hacl.Bignum.Exponentiation\nmodule BM = Hacl.Bignum.Montgomery\nmodule SD = Hacl.Spec.Bignum.Definitions\n\n#reset-options \"--z3rlimit 50 --fuel 0 --ifuel 0\"\n\ninline_for_extraction noextract\nlet bn_check_mod_inv_prime_st (t:limb_t) (len:BN.meta_len t) =\n n:lbignum t len\n -> a:lbignum t len ->\n Stack (limb t)\n (requires fun h -> live h n /\\ live h a /\\ disjoint n a)\n (ensures fun h0 r h1 -> modifies0 h0 h1 /\\\n r == S.bn_check_mod_inv_prime (as_seq h0 n) (as_seq h0 a))\n\n\ninline_for_extraction noextract\nval bn_check_mod_inv_prime: #t:limb_t -> len:BN.meta_len t -> bn_check_mod_inv_prime_st t len\nlet bn_check_mod_inv_prime #t len n a =\n let m0 = BM.bn_check_modulus n in\n let m1 = BN.bn_is_zero_mask len a in\n let m2 = BN.bn_lt_mask len a n in\n m0 &. (lognot m1) &. m2\n\n\ninline_for_extraction noextract\nval bn_mod_inv_prime_n2:\n #t:limb_t\n -> len:BN.meta_len t\n -> n:lbignum t len\n -> res:lbignum t len ->\n Stack unit\n (requires fun h -> 1 < bn_v h n /\\\n live h n /\\ live h res /\\ disjoint res n)\n (ensures fun h0 _ h1 -> modifies (loc res) h0 h1 /\\\n as_seq h1 res == S.bn_mod_inv_prime_n2 (as_seq h0 n))\n\nPartial implementation:\nlet bn_mod_inv_prime_n2 #t len n res =\n \n"
},
{
"role": "assistant",
"content": "let c = BN.bn_sub1 len n (uint #t #SEC 2) res in\nLowStar.Ignore.ignore c;\n()"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: wp: AlgHeap.st_wp a -> Type0\nEffect: Prims.Tot\n\nContext:\nmodule AlgHeap\n\n(* Essentially a copy of AlgForAll but using a heap for the state *)\n\nopen FStar.Tactics.V2\nopen FStar.List.Tot\nopen FStar.Universe\nmodule L = Lattice\nmodule Ghost = FStar.Ghost\nmodule Map = FStar.Map\nmodule T = FStar.Tactics.V2\n\ntype loc = int\ntype state = Map.t loc int\n\ntype empty : Type u#aa =\n\ntype op =\n | Read\n | Write\n | Raise\n | Other of int\n\nassume val other_inp : int -> Type u#0\nlet op_inp : op -> Type u#0 =\n function\n | Read -> unit\n | Write -> state\n | Raise -> exn\n | Other i -> other_inp i\n\nassume val other_out : int -> Type u#0\nlet op_out : op -> Type =\n function\n | Read -> state\n | Write -> unit\n | Raise -> empty\n | Other i -> other_inp i\n\nnoeq\ntype tree0 (a:Type) : Type =\n | Return : a -> tree0 a\n | Op : op:op -> i:(op_inp op) -> k:(op_out op -> tree0 a) -> tree0 a\n\ntype ops = list op\n\nlet rec abides #a (labs:ops) (f : tree0 a) : prop =\n begin match f with\n | Op a i k ->\n mem a labs /\\ (forall o. abides labs (k o))\n | Return _ -> True\n end\n\ntype tree (a:Type u#aa)\n (labs : list u#0 op)\n : Type u#aa\n =\n r:(tree0 a){abides labs r}\n\nlet rec interp_at (l1 l2 : ops) (l : op)\n : Lemma (mem l (l1@l2) == (mem l l1 || mem l l2))\n [SMTPat (mem l (l1@l2))]\n = match l1 with\n | [] -> ()\n | _::l1 -> interp_at l1 l2 l\n\nlet sublist (l1 l2 : ops) = forall x. mem x l1 ==> mem x l2\n\nlet rec sublist_at\n (l1 l2 : ops)\n : Lemma (sublist l1 (l1@l2) /\\ sublist l2 (l1@l2))\n [SMTPatOr [[SMTPat (sublist l1 (l1@l2))];\n [SMTPat (sublist l2 (l1@l2))]]]\n = match l1 with\n | [] -> ()\n | _::l1 -> sublist_at l1 l2\n\nlet sublist_at_self\n (l : ops)\n : Lemma (sublist l (l@l))\n [SMTPat (sublist l (l@l))]\n = ()\n \nlet rec abides_sublist_nopat #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c) /\\ sublist l1 l2)\n (ensures (abides l2) c)\n = match c with\n | Return _ -> ()\n | Op a i k ->\n let sub o : Lemma (abides l2 (k o)) =\n abides_sublist_nopat l1 l2 (k o)\n in\n Classical.forall_intro sub\n\nlet abides_sublist #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c) /\\ sublist l1 l2)\n (ensures (abides l2 c))\n [SMTPat (abides l2 c); SMTPat (sublist l1 l2)]\n = abides_sublist_nopat l1 l2 c\n \nlet abides_at_self #a\n (l : ops)\n (c : tree0 a)\n : Lemma (abides (l@l) c <==> abides l c)\n [SMTPat (abides (l@l) c)]\n = (* Trigger some patterns *)\n assert (sublist l (l@l));\n assert (sublist (l@l) l)\n\nlet abides_app #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c \\/ abides l2 c))\n (ensures (abides (l1@l2) c))\n [SMTPat (abides (l1@l2) c)]\n = sublist_at l1 l2\n\n(* Folding a computation tree *)\nval fold_with (#a #b:_) (#labs : ops)\n (f:tree a labs)\n (v : a -> b)\n (h: (o:op{mem o labs} -> op_inp o -> (op_out o -> b) -> b))\n : b\nlet rec fold_with #a #b #labs f v h =\n match f with\n | Return x -> v x\n | Op act i k ->\n let k' (o : op_out act) : b =\n fold_with #_ #_ #labs (k o) v h\n in\n h act i k'\n\nlet handler_ty_l (o:op) (b:Type) (labs:ops) =\n op_inp o -> (op_out o -> tree b labs) -> tree b labs\n\nlet handler_ty (labs0 : ops) (b:Type) (labs1 : ops) : Type =\n o:op{mem o labs0} -> handler_ty_l o b labs1\n\n(* The most generic handling construct, we use it to implement bind.\nIt is actually just a special case of folding. *)\nval handle_with (#a #b:_) (#labs0 #labs1 : ops)\n (f:tree a labs0)\n (v : a -> tree b labs1)\n (h: handler_ty labs0 b labs1)\n : tree b labs1\nlet handle_with f v h = fold_with f v h\n\nlet return (a:Type) (x:a)\n : tree a []\n = Return x\n\nlet bind (a b : Type)\n (#labs1 #labs2 : ops)\n (c : tree a labs1)\n (f : (x:a -> tree b labs2))\n : Tot (tree b (labs1@labs2))\n = handle_with #_ #_ #labs1 #(labs1@labs2) c f (fun act i k -> Op act i k)\n \nlet subcomp (a:Type)\n (labs1 labs2 : ops)\n (f : tree a labs1)\n : Pure (tree a labs2)\n (requires (sublist labs1 labs2))\n (ensures (fun _ -> True))\n = f\n\nlet if_then_else\n (a : Type)\n (labs1 labs2 : ops)\n (f : tree a labs1)\n (g : tree a labs2)\n (p : bool)\n : Type\n = tree a (labs1@labs2)\n\nlet _get : tree state [Read] = Op Read () Return\n\nlet _put (s:state) : tree unit [Write] = Op Write s Return\n\ntotal // need this for catch!!\nreifiable\nreflectable\neffect {\n Alg (a:Type) (_:ops)\n with {repr = tree; return; bind; subcomp; if_then_else}\n}\n\nlet get () : Alg state [Read] =\n Alg?.reflect _get\n\nlet lift_pure_eff\n (a:Type)\n (wp : pure_wp a)\n (f : unit -> PURE a wp)\n : Pure (tree a [])\n (requires (wp (fun _ -> True)))\n (ensures (fun _ -> True))\n = FStar.Monotonic.Pure.elim_pure_wp_monotonicity wp;\n Return (f ())\n\nsub_effect PURE ~> Alg = lift_pure_eff\n\nlet put (s:state) : Alg unit [Write] =\n Alg?.reflect (_put s)\n\nlet raise #a (e:exn) : Alg a [Raise] = \n Alg?.reflect (Op Raise e (fun e -> match e with))\n // funnily enough, the version below also succeeds from concluding\n // a==empty under the lambda since the context becomes inconsistent\n //Alg?.reflect (Op Raise e Return\n\ntype rwtree a = tree a [Read;Write]\n\nlet tbind : #a:_ -> #b:_ -> rwtree a -> (a -> rwtree b) -> rwtree b = fun c f -> bind _ _ c f\n\nlet st_wp0 (a:Type) : Type = state -> (a & state -> Type0) -> Type0\n\nlet st_monotonic #a (w : st_wp0 a) : Type0 =\n //forall s0 p1 p2. (forall r. p1 r ==> p2 r) ==> w s0 p1 ==> w s0 p2\n // ^ this version seems to be less SMT-friendly\n forall s0 p1 p2. (forall x s1. p1 (x, s1) ==> p2 (x, s1)) ==> w s0 p1 ==> w s0 p2\n\nlet st_wp (a:Type) = wp:st_wp0 a{st_monotonic wp}\n\nunfold\nlet return_wp #a x : st_wp a = fun s0 p -> p (x, s0)\n\nunfold\nlet bind_wp #a #b (w : st_wp a) (wf : a -> st_wp b)\n : st_wp b\n = fun s0 p -> w s0 (fun (y, s1) -> wf y s1 p)\n\nunfold\nlet read_wp : st_wp state = fun s0 p -> p (s0, s0)\n\nunfold\nlet write_wp : state -> st_wp unit = fun s _ p -> p ((), s)\n\n(* Also doable with handlers *)\nlet rec interp_as_wp #a (t : rwtree a) : st_wp a =\n match t with\n | Return x -> return_wp x\n | Op Read _ k ->\n bind_wp read_wp (fun s -> interp_as_wp (k s))\n | Op Write s k ->\n bind_wp (write_wp s) (fun (o:unit) -> interp_as_wp (k o))\n\nlet rec interp_rdwr_tree #a (t : tree a [Read;Write]) (s:state) : Tot (a & state) =\n match t with\n | Return x -> (x, s)\n | Op Read _ k ->\n interp_rdwr_tree (k s) s\n | Op Write s k ->\n interp_rdwr_tree (k ()) s\n\nlet interp_as_fun #a (t : rwtree a) : (state -> a & state) =\n interp_rdwr_tree t\n\n(* Bug: defining this as a FStar.Preorder.preorder\ncauses stupid failures ahead *)\nval stronger : (#a:Type) -> st_wp a -> st_wp a -> Type0\nlet stronger w1 w2 = forall p s. w1 p s ==> w2 p s\n\nlet equiv #a (w1 w2 : st_wp a) = w1 `stronger` w2 /\\ w2 `stronger` w1\n\nlet (<<=) = stronger\n\nval interp_ret (#a:Type) (x:a) : Lemma (return_wp x `stronger` interp_as_wp (Return x))\nlet interp_ret x = ()\n\nPartial implementation:\nlet wp_is_monotonic #a (wp: st_wp a) : Type0 =\n \n"
},
{
"role": "assistant",
"content": "forall p1 p2 s0. (forall x s1. p1 (x, s1) ==> p2 (x, s1)) ==> wp s0 p1 ==> wp s0 p2"
}
] |
[
{
"role": "user",
"content": "Complete this F* function with a correct proof.\n\nType: a: Type -> b: Type -> c: AlgHeap.tree a labs1 -> f: (x: a -> AlgHeap.tree b labs2)\n -> AlgHeap.tree b (labs1 @ labs2)\nEffect: Prims.Tot\n\nContext:\nmodule AlgHeap\n\n(* Essentially a copy of AlgForAll but using a heap for the state *)\n\nopen FStar.Tactics.V2\nopen FStar.List.Tot\nopen FStar.Universe\nmodule L = Lattice\nmodule Ghost = FStar.Ghost\nmodule Map = FStar.Map\nmodule T = FStar.Tactics.V2\n\ntype loc = int\ntype state = Map.t loc int\n\ntype empty : Type u#aa =\n\ntype op =\n | Read\n | Write\n | Raise\n | Other of int\n\nassume val other_inp : int -> Type u#0\nlet op_inp : op -> Type u#0 =\n function\n | Read -> unit\n | Write -> state\n | Raise -> exn\n | Other i -> other_inp i\n\nassume val other_out : int -> Type u#0\nlet op_out : op -> Type =\n function\n | Read -> state\n | Write -> unit\n | Raise -> empty\n | Other i -> other_inp i\n\nnoeq\ntype tree0 (a:Type) : Type =\n | Return : a -> tree0 a\n | Op : op:op -> i:(op_inp op) -> k:(op_out op -> tree0 a) -> tree0 a\n\ntype ops = list op\n\nlet rec abides #a (labs:ops) (f : tree0 a) : prop =\n begin match f with\n | Op a i k ->\n mem a labs /\\ (forall o. abides labs (k o))\n | Return _ -> True\n end\n\ntype tree (a:Type u#aa)\n (labs : list u#0 op)\n : Type u#aa\n =\n r:(tree0 a){abides labs r}\n\nlet rec interp_at (l1 l2 : ops) (l : op)\n : Lemma (mem l (l1@l2) == (mem l l1 || mem l l2))\n [SMTPat (mem l (l1@l2))]\n = match l1 with\n | [] -> ()\n | _::l1 -> interp_at l1 l2 l\n\nlet sublist (l1 l2 : ops) = forall x. mem x l1 ==> mem x l2\n\nlet rec sublist_at\n (l1 l2 : ops)\n : Lemma (sublist l1 (l1@l2) /\\ sublist l2 (l1@l2))\n [SMTPatOr [[SMTPat (sublist l1 (l1@l2))];\n [SMTPat (sublist l2 (l1@l2))]]]\n = match l1 with\n | [] -> ()\n | _::l1 -> sublist_at l1 l2\n\nlet sublist_at_self\n (l : ops)\n : Lemma (sublist l (l@l))\n [SMTPat (sublist l (l@l))]\n = ()\n \nlet rec abides_sublist_nopat #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c) /\\ sublist l1 l2)\n (ensures (abides l2) c)\n = match c with\n | Return _ -> ()\n | Op a i k ->\n let sub o : Lemma (abides l2 (k o)) =\n abides_sublist_nopat l1 l2 (k o)\n in\n Classical.forall_intro sub\n\nlet abides_sublist #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c) /\\ sublist l1 l2)\n (ensures (abides l2 c))\n [SMTPat (abides l2 c); SMTPat (sublist l1 l2)]\n = abides_sublist_nopat l1 l2 c\n \nlet abides_at_self #a\n (l : ops)\n (c : tree0 a)\n : Lemma (abides (l@l) c <==> abides l c)\n [SMTPat (abides (l@l) c)]\n = (* Trigger some patterns *)\n assert (sublist l (l@l));\n assert (sublist (l@l) l)\n\nlet abides_app #a (l1 l2 : ops) (c : tree0 a)\n : Lemma (requires (abides l1 c \\/ abides l2 c))\n (ensures (abides (l1@l2) c))\n [SMTPat (abides (l1@l2) c)]\n = sublist_at l1 l2\n\n(* Folding a computation tree *)\nval fold_with (#a #b:_) (#labs : ops)\n (f:tree a labs)\n (v : a -> b)\n (h: (o:op{mem o labs} -> op_inp o -> (op_out o -> b) -> b))\n : b\nlet rec fold_with #a #b #labs f v h =\n match f with\n | Return x -> v x\n | Op act i k ->\n let k' (o : op_out act) : b =\n fold_with #_ #_ #labs (k o) v h\n in\n h act i k'\n\nlet handler_ty_l (o:op) (b:Type) (labs:ops) =\n op_inp o -> (op_out o -> tree b labs) -> tree b labs\n\nlet handler_ty (labs0 : ops) (b:Type) (labs1 : ops) : Type =\n o:op{mem o labs0} -> handler_ty_l o b labs1\n\n(* The most generic handling construct, we use it to implement bind.\nIt is actually just a special case of folding. *)\nval handle_with (#a #b:_) (#labs0 #labs1 : ops)\n (f:tree a labs0)\n (v : a -> tree b labs1)\n (h: handler_ty labs0 b labs1)\n : tree b labs1\nlet handle_with f v h = fold_with f v h\n\nlet return (a:Type) (x:a)\n : tree a []\n = Return x\n\nPartial implementation:\nlet bind (a b: Type) (#labs1 #labs2: ops) (c: tree a labs1) (f: (x: a -> tree b labs2))\n : Tot (tree b (labs1 @ labs2)) =\n \n"
},
{
"role": "assistant",
"content": "handle_with #_ #_ #labs1 #(labs1 @ labs2) c f (fun act i k -> Op act i k)"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.