xref: /XiangShan/src/main/scala/xiangshan/cache/mmu/PageTableWalker.scala (revision 92e3bfefd90c9c71e8a65e5955c61ee13765fb9d)
16d5ddbceSLemover/***************************************************************************************
26d5ddbceSLemover* Copyright (c) 2020-2021 Institute of Computing Technology, Chinese Academy of Sciences
3f320e0f0SYinan Xu* Copyright (c) 2020-2021 Peng Cheng Laboratory
46d5ddbceSLemover*
56d5ddbceSLemover* XiangShan is licensed under Mulan PSL v2.
66d5ddbceSLemover* You can use this software according to the terms and conditions of the Mulan PSL v2.
76d5ddbceSLemover* You may obtain a copy of Mulan PSL v2 at:
86d5ddbceSLemover*          http://license.coscl.org.cn/MulanPSL2
96d5ddbceSLemover*
106d5ddbceSLemover* THIS SOFTWARE IS PROVIDED ON AN "AS IS" BASIS, WITHOUT WARRANTIES OF ANY KIND,
116d5ddbceSLemover* EITHER EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO NON-INFRINGEMENT,
126d5ddbceSLemover* MERCHANTABILITY OR FIT FOR A PARTICULAR PURPOSE.
136d5ddbceSLemover*
146d5ddbceSLemover* See the Mulan PSL v2 for more details.
156d5ddbceSLemover***************************************************************************************/
166d5ddbceSLemover
176d5ddbceSLemoverpackage xiangshan.cache.mmu
186d5ddbceSLemover
196d5ddbceSLemoverimport chipsalliance.rocketchip.config.Parameters
206d5ddbceSLemoverimport chisel3._
216d5ddbceSLemoverimport chisel3.util._
22b848eea5SLemoverimport chisel3.internal.naming.chiselName
236d5ddbceSLemoverimport xiangshan._
246d5ddbceSLemoverimport xiangshan.cache.{HasDCacheParameters, MemoryOpConstants}
256d5ddbceSLemoverimport utils._
266d5ddbceSLemoverimport freechips.rocketchip.diplomacy.{LazyModule, LazyModuleImp}
276d5ddbceSLemoverimport freechips.rocketchip.tilelink._
28b6982e83SLemoverimport xiangshan.backend.fu.{PMPReqBundle, PMPRespBundle}
296d5ddbceSLemover
30*92e3bfefSLemover/** Page Table Walk is divided into two parts
31*92e3bfefSLemover  * One,   PTW: page walk for pde, except for leaf entries, one by one
32*92e3bfefSLemover  * Two, LLPTW: page walk for pte, only the leaf entries(4KB), in parallel
336d5ddbceSLemover  */
34*92e3bfefSLemover
35*92e3bfefSLemover
36*92e3bfefSLemover/** PTW : page table walker
37*92e3bfefSLemover  * a finite state machine
38*92e3bfefSLemover  * only take 1GB and 2MB page walks
39*92e3bfefSLemover  * or in other words, except the last level(leaf)
40*92e3bfefSLemover  **/
41*92e3bfefSLemoverclass PTWIO()(implicit p: Parameters) extends MMUIOBaseBundle with HasPtwConst {
426d5ddbceSLemover  val req = Flipped(DecoupledIO(new Bundle {
4345f497a4Shappy-lx    val req_info = new L2TlbInnerBundle()
446d5ddbceSLemover    val l1Hit = Bool()
456d5ddbceSLemover    val ppn = UInt(ppnLen.W)
466d5ddbceSLemover  }))
476d5ddbceSLemover  val resp = DecoupledIO(new Bundle {
48bc063562SLemover    val source = UInt(bSourceWidth.W)
496d5ddbceSLemover    val resp = new PtwResp
506d5ddbceSLemover  })
516d5ddbceSLemover
52*92e3bfefSLemover  val llptw = DecoupledIO(new LLPTWInBundle())
53cc5a5f22SLemover
546d5ddbceSLemover  val mem = new Bundle {
55b848eea5SLemover    val req = DecoupledIO(new L2TlbMemReqBundle())
565854c1edSLemover    val resp = Flipped(ValidIO(UInt(XLEN.W)))
57cc5a5f22SLemover    val mask = Input(Bool())
586d5ddbceSLemover  }
59b6982e83SLemover  val pmp = new Bundle {
60b6982e83SLemover    val req = ValidIO(new PMPReqBundle())
61b6982e83SLemover    val resp = Flipped(new PMPRespBundle())
62b6982e83SLemover  }
636d5ddbceSLemover
646d5ddbceSLemover  val refill = Output(new Bundle {
6545f497a4Shappy-lx    val req_info = new L2TlbInnerBundle()
666d5ddbceSLemover    val level = UInt(log2Up(Level).W)
676d5ddbceSLemover  })
686d5ddbceSLemover}
696d5ddbceSLemover
70b848eea5SLemover@chiselName
71*92e3bfefSLemoverclass PTW()(implicit p: Parameters) extends XSModule with HasPtwConst with HasPerfEvents {
72*92e3bfefSLemover  val io = IO(new PTWIO)
736d5ddbceSLemover
746d5ddbceSLemover  val sfence = io.sfence
756d5ddbceSLemover  val mem = io.mem
766d5ddbceSLemover  val satp = io.csr.satp
7745f497a4Shappy-lx  val flush = io.sfence.valid || io.csr.satp.changed
786d5ddbceSLemover
79b6982e83SLemover  val s_idle :: s_addr_check :: s_mem_req :: s_mem_resp :: s_check_pte :: Nil = Enum(5)
806d5ddbceSLemover  val state = RegInit(s_idle)
816d5ddbceSLemover  val level = RegInit(0.U(log2Up(Level).W))
82b6982e83SLemover  val af_level = RegInit(0.U(log2Up(Level).W)) // access fault return this level
836d5ddbceSLemover  val ppn = Reg(UInt(ppnLen.W))
846d5ddbceSLemover  val vpn = Reg(UInt(vpnLen.W))
856d5ddbceSLemover  val levelNext = level + 1.U
866d5ddbceSLemover  val l1Hit = Reg(Bool())
875854c1edSLemover  val memPte = mem.resp.bits.asTypeOf(new PteBundle().cloneType)
88cc5a5f22SLemover  io.req.ready := state === s_idle
896d5ddbceSLemover
90b6982e83SLemover  val finish = WireInit(false.B)
91b6982e83SLemover  val sent_to_pmp = state === s_addr_check || (state === s_check_pte && !finish)
92ca2f90a6SLemover  val accessFault = RegEnable(io.pmp.resp.ld || io.pmp.resp.mmio, sent_to_pmp)
93b6982e83SLemover  val pageFault = memPte.isPf(level)
946d5ddbceSLemover  switch (state) {
956d5ddbceSLemover    is (s_idle) {
966d5ddbceSLemover      when (io.req.fire()) {
976d5ddbceSLemover        val req = io.req.bits
98b6982e83SLemover        state := s_addr_check
99cc5a5f22SLemover        level := Mux(req.l1Hit, 1.U, 0.U)
100b6982e83SLemover        af_level := Mux(req.l1Hit, 1.U, 0.U)
101cc5a5f22SLemover        ppn := Mux(req.l1Hit, io.req.bits.ppn, satp.ppn)
10245f497a4Shappy-lx        vpn := io.req.bits.req_info.vpn
1036d5ddbceSLemover        l1Hit := req.l1Hit
104b6982e83SLemover        accessFault := false.B
1056d5ddbceSLemover      }
1066d5ddbceSLemover    }
1076d5ddbceSLemover
108b6982e83SLemover    is (s_addr_check) {
109b6982e83SLemover      state := s_mem_req
110b6982e83SLemover    }
111b6982e83SLemover
1126d5ddbceSLemover    is (s_mem_req) {
1136d5ddbceSLemover      when (mem.req.fire()) {
1146d5ddbceSLemover        state := s_mem_resp
1156d5ddbceSLemover      }
116b6982e83SLemover      when (accessFault) {
117b6982e83SLemover        state := s_check_pte
118b6982e83SLemover      }
1196d5ddbceSLemover    }
1206d5ddbceSLemover
1216d5ddbceSLemover    is (s_mem_resp) {
1226d5ddbceSLemover      when(mem.resp.fire()) {
123cc5a5f22SLemover        state := s_check_pte
124b6982e83SLemover        af_level := af_level + 1.U
1256d5ddbceSLemover      }
1266d5ddbceSLemover    }
1276d5ddbceSLemover
128cc5a5f22SLemover    is (s_check_pte) {
1292b773508SZhangZifei      when (io.resp.valid) { // find pte already or accessFault (mentioned below)
1306d5ddbceSLemover        when (io.resp.fire()) {
1316d5ddbceSLemover          state := s_idle
1326d5ddbceSLemover        }
133b6982e83SLemover        finish := true.B
134*92e3bfefSLemover      }.elsewhen(io.llptw.valid) { // the next level is pte, go to miss queue
135*92e3bfefSLemover        when (io.llptw.fire()) {
136cc5a5f22SLemover          state := s_idle
137cc5a5f22SLemover        }
138b6982e83SLemover        finish := true.B
1392b773508SZhangZifei      } otherwise { // go to next level, access the memory, need pmp check first
1402b773508SZhangZifei        when (io.pmp.resp.ld) { // pmp check failed, raise access-fault
1412b773508SZhangZifei          // do nothing, RegNext the pmp check result and do it later (mentioned above)
1422b773508SZhangZifei        }.otherwise { // go to next level.
1432b773508SZhangZifei          assert(level === 0.U)
144b6982e83SLemover          level := levelNext
145b6982e83SLemover          state := s_mem_req
146cc5a5f22SLemover        }
147cc5a5f22SLemover      }
1486d5ddbceSLemover    }
1496d5ddbceSLemover  }
1506d5ddbceSLemover
1516d5ddbceSLemover  when (sfence.valid) {
1526d5ddbceSLemover    state := s_idle
153b6982e83SLemover    accessFault := false.B
1546d5ddbceSLemover  }
1556d5ddbceSLemover
156b6982e83SLemover  // memPte is valid when at s_check_pte. when mem.resp.fire, it's not ready.
157cc5a5f22SLemover  val is_pte = memPte.isLeaf() || memPte.isPf(level)
158cc5a5f22SLemover  val find_pte = is_pte
159cc5a5f22SLemover  val to_find_pte = level === 1.U && !is_pte
16045f497a4Shappy-lx  val source = RegEnable(io.req.bits.req_info.source, io.req.fire())
161b6982e83SLemover  io.resp.valid := state === s_check_pte && (find_pte || accessFault)
162cc5a5f22SLemover  io.resp.bits.source := source
16345f497a4Shappy-lx  io.resp.bits.resp.apply(pageFault && !accessFault, accessFault, Mux(accessFault, af_level, level), memPte, vpn, satp.asid)
164cc5a5f22SLemover
165*92e3bfefSLemover  io.llptw.valid := state === s_check_pte && to_find_pte && !accessFault
166*92e3bfefSLemover  io.llptw.bits.req_info.source := source
167*92e3bfefSLemover  io.llptw.bits.req_info.vpn := vpn
168*92e3bfefSLemover  io.llptw.bits.ppn := memPte.ppn
169cc5a5f22SLemover
170cc5a5f22SLemover  assert(level =/= 2.U || level =/= 3.U)
1716d5ddbceSLemover
1726d5ddbceSLemover  val l1addr = MakeAddr(satp.ppn, getVpnn(vpn, 2))
173cc5a5f22SLemover  val l2addr = MakeAddr(Mux(l1Hit, ppn, memPte.ppn), getVpnn(vpn, 1))
174b6982e83SLemover  val mem_addr = Mux(af_level === 0.U, l1addr, l2addr)
175b6982e83SLemover  io.pmp.req.valid := DontCare // samecycle, do not use valid
176b6982e83SLemover  io.pmp.req.bits.addr := mem_addr
177b6982e83SLemover  io.pmp.req.bits.size := 3.U // TODO: fix it
178b6982e83SLemover  io.pmp.req.bits.cmd := TlbCmd.read
179b6982e83SLemover
180b6982e83SLemover  mem.req.valid := state === s_mem_req && !io.mem.mask && !accessFault
181b6982e83SLemover  mem.req.bits.addr := mem_addr
182bc063562SLemover  mem.req.bits.id := FsmReqID.U(bMemID.W)
1836d5ddbceSLemover
18445f497a4Shappy-lx  io.refill.req_info.vpn := vpn
1856d5ddbceSLemover  io.refill.level := level
18645f497a4Shappy-lx  io.refill.req_info.source := source
1876d5ddbceSLemover
188*92e3bfefSLemover  XSDebug(p"[ptw] state:${state} level:${level} notFound:${pageFault}\n")
1896d5ddbceSLemover
1906d5ddbceSLemover  // perf
1916d5ddbceSLemover  XSPerfAccumulate("fsm_count", io.req.fire())
1926d5ddbceSLemover  for (i <- 0 until PtwWidth) {
19345f497a4Shappy-lx    XSPerfAccumulate(s"fsm_count_source${i}", io.req.fire() && io.req.bits.req_info.source === i.U)
1946d5ddbceSLemover  }
1956d5ddbceSLemover  XSPerfAccumulate("fsm_busy", state =/= s_idle)
1966d5ddbceSLemover  XSPerfAccumulate("fsm_idle", state === s_idle)
1976d5ddbceSLemover  XSPerfAccumulate("resp_blocked", io.resp.valid && !io.resp.ready)
1986d5ddbceSLemover  XSPerfAccumulate("mem_count", mem.req.fire())
1996d5ddbceSLemover  XSPerfAccumulate("mem_cycle", BoolStopWatch(mem.req.fire, mem.resp.fire(), true))
2006d5ddbceSLemover  XSPerfAccumulate("mem_blocked", mem.req.valid && !mem.req.ready)
201cc5a5f22SLemover
2029bd9cdfaSLemover  TimeOutAssert(state =/= s_idle, timeOutThreshold, "page table walker time out")
203cd365d4cSrvcoresjw
204cd365d4cSrvcoresjw  val perfEvents = Seq(
205cd365d4cSrvcoresjw    ("fsm_count         ", io.req.fire()                                     ),
206cd365d4cSrvcoresjw    ("fsm_busy          ", state =/= s_idle                                  ),
207cd365d4cSrvcoresjw    ("fsm_idle          ", state === s_idle                                  ),
208cd365d4cSrvcoresjw    ("resp_blocked      ", io.resp.valid && !io.resp.ready                   ),
209cd365d4cSrvcoresjw    ("mem_count         ", mem.req.fire()                                    ),
210cd365d4cSrvcoresjw    ("mem_cycle         ", BoolStopWatch(mem.req.fire, mem.resp.fire(), true)),
211cd365d4cSrvcoresjw    ("mem_blocked       ", mem.req.valid && !mem.req.ready                   ),
212cd365d4cSrvcoresjw  )
2131ca0e4f3SYinan Xu  generatePerfEvent()
2146d5ddbceSLemover}
215*92e3bfefSLemover
216*92e3bfefSLemover/*========================= LLPTW ==============================*/
217*92e3bfefSLemover
218*92e3bfefSLemover/** LLPTW : Last Level Page Table Walker
219*92e3bfefSLemover  * the page walker that only takes 4KB(last level) page walk.
220*92e3bfefSLemover  **/
221*92e3bfefSLemover
222*92e3bfefSLemoverclass LLPTWInBundle(implicit p: Parameters) extends XSBundle with HasPtwConst {
223*92e3bfefSLemover  val req_info = Output(new L2TlbInnerBundle())
224*92e3bfefSLemover  val ppn = Output(UInt(PAddrBits.W))
225*92e3bfefSLemover}
226*92e3bfefSLemover
227*92e3bfefSLemoverclass LLPTWIO(implicit p: Parameters) extends MMUIOBaseBundle with HasPtwConst {
228*92e3bfefSLemover  val in = Flipped(DecoupledIO(new LLPTWInBundle()))
229*92e3bfefSLemover  val out = DecoupledIO(new Bundle {
230*92e3bfefSLemover    val req_info = Output(new L2TlbInnerBundle())
231*92e3bfefSLemover    val id = Output(UInt(bMemID.W))
232*92e3bfefSLemover    val af = Output(Bool())
233*92e3bfefSLemover  })
234*92e3bfefSLemover  val mem = new Bundle {
235*92e3bfefSLemover    val req = DecoupledIO(new L2TlbMemReqBundle())
236*92e3bfefSLemover    val resp = Flipped(Valid(new Bundle {
237*92e3bfefSLemover      val id = Output(UInt(log2Up(l2tlbParams.llptwsize).W))
238*92e3bfefSLemover    }))
239*92e3bfefSLemover    val enq_ptr = Output(UInt(log2Ceil(l2tlbParams.llptwsize).W))
240*92e3bfefSLemover    val buffer_it = Output(Vec(l2tlbParams.llptwsize, Bool()))
241*92e3bfefSLemover    val refill = Output(new L2TlbInnerBundle())
242*92e3bfefSLemover    val req_mask = Input(Vec(l2tlbParams.llptwsize, Bool()))
243*92e3bfefSLemover  }
244*92e3bfefSLemover  val pmp = new Bundle {
245*92e3bfefSLemover    val req = Valid(new PMPReqBundle())
246*92e3bfefSLemover    val resp = Flipped(new PMPRespBundle())
247*92e3bfefSLemover  }
248*92e3bfefSLemover}
249*92e3bfefSLemover
250*92e3bfefSLemoverclass LLPTWEntry(implicit p: Parameters) extends XSBundle with HasPtwConst {
251*92e3bfefSLemover  val req_info = new L2TlbInnerBundle()
252*92e3bfefSLemover  val ppn = UInt(ppnLen.W)
253*92e3bfefSLemover  val wait_id = UInt(log2Up(l2tlbParams.llptwsize).W)
254*92e3bfefSLemover  val af = Bool()
255*92e3bfefSLemover}
256*92e3bfefSLemover
257*92e3bfefSLemover
258*92e3bfefSLemover@chiselName
259*92e3bfefSLemoverclass LLPTW(implicit p: Parameters) extends XSModule with HasPtwConst with HasPerfEvents {
260*92e3bfefSLemover  val io = IO(new LLPTWIO())
261*92e3bfefSLemover
262*92e3bfefSLemover  val entries = Reg(Vec(l2tlbParams.llptwsize, new LLPTWEntry()))
263*92e3bfefSLemover  val state_idle :: state_addr_check :: state_mem_req :: state_mem_waiting :: state_mem_out :: Nil = Enum(5)
264*92e3bfefSLemover  val state = RegInit(VecInit(Seq.fill(l2tlbParams.llptwsize)(state_idle)))
265*92e3bfefSLemover  val is_emptys = state.map(_ === state_idle)
266*92e3bfefSLemover  val is_mems = state.map(_ === state_mem_req)
267*92e3bfefSLemover  val is_waiting = state.map(_ === state_mem_waiting)
268*92e3bfefSLemover  val is_having = state.map(_ === state_mem_out)
269*92e3bfefSLemover
270*92e3bfefSLemover  val full = !ParallelOR(is_emptys).asBool()
271*92e3bfefSLemover  val enq_ptr = ParallelPriorityEncoder(is_emptys)
272*92e3bfefSLemover
273*92e3bfefSLemover  val mem_ptr = ParallelPriorityEncoder(is_having)
274*92e3bfefSLemover  val mem_arb = Module(new RRArbiter(new LLPTWEntry(), l2tlbParams.llptwsize))
275*92e3bfefSLemover  for (i <- 0 until l2tlbParams.llptwsize) {
276*92e3bfefSLemover    mem_arb.io.in(i).bits := entries(i)
277*92e3bfefSLemover    mem_arb.io.in(i).valid := is_mems(i) && !io.mem.req_mask(i)
278*92e3bfefSLemover  }
279*92e3bfefSLemover
280*92e3bfefSLemover  // duplicate req
281*92e3bfefSLemover  // to_wait: wait for the last to access mem, set to mem_resp
282*92e3bfefSLemover  // to_cache: the last is back just right now, set to mem_cache
283*92e3bfefSLemover  def dup(vpn1: UInt, vpn2: UInt): Bool = {
284*92e3bfefSLemover    dropL3SectorBits(vpn1) === dropL3SectorBits(vpn2)
285*92e3bfefSLemover  }
286*92e3bfefSLemover  val dup_vec = state.indices.map(i =>
287*92e3bfefSLemover    dup(io.in.bits.req_info.vpn, entries(i).req_info.vpn)
288*92e3bfefSLemover  )
289*92e3bfefSLemover  val dup_req_fire = mem_arb.io.out.fire() && dup(io.in.bits.req_info.vpn, mem_arb.io.out.bits.req_info.vpn) // dup with the req fire entry
290*92e3bfefSLemover  val dup_vec_wait = dup_vec.zip(is_waiting).map{case (d, w) => d && w} // dup with "mem_waiting" entres, sending mem req already
291*92e3bfefSLemover  val dup_vec_having = dup_vec.zipWithIndex.map{case (d, i) => d && is_having(i)} // dup with the "mem_out" entry recv the data just now
292*92e3bfefSLemover  val wait_id = Mux(dup_req_fire, mem_arb.io.chosen, ParallelMux(dup_vec_wait zip entries.map(_.wait_id)))
293*92e3bfefSLemover  val dup_wait_resp = io.mem.resp.fire() && VecInit(dup_vec_wait)(io.mem.resp.bits.id) // dup with the entry that data coming next cycle
294*92e3bfefSLemover  val to_wait = Cat(dup_vec_wait).orR || dup_req_fire
295*92e3bfefSLemover  val to_mem_out = dup_wait_resp
296*92e3bfefSLemover  val to_cache_low = Cat(dup_vec_having).orR
297*92e3bfefSLemover  assert(RegNext(!(dup_req_fire && Cat(dup_vec_wait).orR), init = true.B), "mem req but some entries already waiting, should not happed")
298*92e3bfefSLemover
299*92e3bfefSLemover  val mem_resp_hit = RegInit(VecInit(Seq.fill(l2tlbParams.llptwsize)(false.B)))
300*92e3bfefSLemover  val enq_state = Mux(to_mem_out, state_mem_out, // same to the blew, but the mem resp now
301*92e3bfefSLemover    Mux(to_wait, state_mem_waiting, state_addr_check))
302*92e3bfefSLemover  when (io.in.fire()) {
303*92e3bfefSLemover    // if prefetch req does not need mem access, just give it up.
304*92e3bfefSLemover    // so there will be at most 1 + FilterSize entries that needs re-access page cache
305*92e3bfefSLemover    // so 2 + FilterSize is enough to avoid dead-lock
306*92e3bfefSLemover    state(enq_ptr) := Mux(from_pre(io.in.bits.req_info.source) && enq_state =/= state_addr_check, state_idle, enq_state)
307*92e3bfefSLemover    entries(enq_ptr).req_info := io.in.bits.req_info
308*92e3bfefSLemover    entries(enq_ptr).ppn := io.in.bits.ppn
309*92e3bfefSLemover    entries(enq_ptr).wait_id := Mux(to_wait, wait_id, enq_ptr)
310*92e3bfefSLemover    entries(enq_ptr).af := false.B
311*92e3bfefSLemover    mem_resp_hit(enq_ptr) := to_mem_out
312*92e3bfefSLemover  }
313*92e3bfefSLemover  when (mem_arb.io.out.fire()) {
314*92e3bfefSLemover    for (i <- state.indices) {
315*92e3bfefSLemover      when (state(i) =/= state_idle && dup(entries(i).req_info.vpn, mem_arb.io.out.bits.req_info.vpn)) {
316*92e3bfefSLemover        // NOTE: "dup enq set state to mem_wait" -> "sending req set other dup entries to mem_wait"
317*92e3bfefSLemover        state(i) := state_mem_waiting
318*92e3bfefSLemover        entries(i).wait_id := mem_arb.io.chosen
319*92e3bfefSLemover      }
320*92e3bfefSLemover    }
321*92e3bfefSLemover  }
322*92e3bfefSLemover  when (io.mem.resp.fire()) {
323*92e3bfefSLemover    state.indices.map{i =>
324*92e3bfefSLemover      when (state(i) === state_mem_waiting && io.mem.resp.bits.id === entries(i).wait_id) {
325*92e3bfefSLemover        state(i) := state_mem_out
326*92e3bfefSLemover        mem_resp_hit(i) := true.B
327*92e3bfefSLemover      }
328*92e3bfefSLemover    }
329*92e3bfefSLemover  }
330*92e3bfefSLemover  when (io.out.fire()) {
331*92e3bfefSLemover    assert(state(mem_ptr) === state_mem_out)
332*92e3bfefSLemover    state(mem_ptr) := state_idle
333*92e3bfefSLemover  }
334*92e3bfefSLemover  mem_resp_hit.map(a => when (a) { a := false.B } )
335*92e3bfefSLemover
336*92e3bfefSLemover  val enq_ptr_reg = RegNext(enq_ptr)
337*92e3bfefSLemover
338*92e3bfefSLemover  io.pmp.req.valid := RegNext(enq_state === state_addr_check)
339*92e3bfefSLemover  io.pmp.req.bits.addr := MakeAddr(entries(enq_ptr_reg).ppn, getVpnn(entries(enq_ptr_reg).req_info.vpn, 0))
340*92e3bfefSLemover  io.pmp.req.bits.cmd := TlbCmd.read
341*92e3bfefSLemover  io.pmp.req.bits.size := 3.U // TODO: fix it
342*92e3bfefSLemover  val pmp_resp_valid = io.pmp.req.valid // same cycle
343*92e3bfefSLemover  when (pmp_resp_valid && (state(enq_ptr_reg) === state_addr_check) &&
344*92e3bfefSLemover    !(mem_arb.io.out.fire && dup(entries(enq_ptr_reg).req_info.vpn, mem_arb.io.out.bits.req_info.vpn))) {
345*92e3bfefSLemover    // NOTE: when pmp resp but state is not addr check, then the entry is dup with other entry, the state was changed before
346*92e3bfefSLemover    //       when dup with the req-ing entry, set to mem_waiting (above codes), and the ld must be false, so dontcare
347*92e3bfefSLemover    val accessFault = io.pmp.resp.ld || io.pmp.resp.mmio
348*92e3bfefSLemover    entries(enq_ptr_reg).af := accessFault
349*92e3bfefSLemover    state(enq_ptr_reg) := Mux(accessFault, state_mem_out, state_mem_req)
350*92e3bfefSLemover  }
351*92e3bfefSLemover
352*92e3bfefSLemover  val flush = io.sfence.valid || io.csr.satp.changed
353*92e3bfefSLemover  when (flush) {
354*92e3bfefSLemover    state.map(_ := state_idle)
355*92e3bfefSLemover  }
356*92e3bfefSLemover
357*92e3bfefSLemover  io.in.ready := !full
358*92e3bfefSLemover
359*92e3bfefSLemover  io.out.valid := ParallelOR(is_having).asBool()
360*92e3bfefSLemover  io.out.bits.req_info := entries(mem_ptr).req_info
361*92e3bfefSLemover  io.out.bits.id := mem_ptr
362*92e3bfefSLemover  io.out.bits.af := entries(mem_ptr).af
363*92e3bfefSLemover
364*92e3bfefSLemover  io.mem.req.valid := mem_arb.io.out.valid && !flush
365*92e3bfefSLemover  io.mem.req.bits.addr := MakeAddr(mem_arb.io.out.bits.ppn, getVpnn(mem_arb.io.out.bits.req_info.vpn, 0))
366*92e3bfefSLemover  io.mem.req.bits.id := mem_arb.io.chosen
367*92e3bfefSLemover  mem_arb.io.out.ready := io.mem.req.ready
368*92e3bfefSLemover  io.mem.refill := entries(RegNext(io.mem.resp.bits.id(log2Up(l2tlbParams.llptwsize)-1, 0))).req_info
369*92e3bfefSLemover  io.mem.buffer_it := mem_resp_hit
370*92e3bfefSLemover  io.mem.enq_ptr := enq_ptr
371*92e3bfefSLemover
372*92e3bfefSLemover  XSPerfAccumulate("llptw_in_count", io.in.fire())
373*92e3bfefSLemover  XSPerfAccumulate("llptw_in_block", io.in.valid && !io.in.ready)
374*92e3bfefSLemover  for (i <- 0 until 7) {
375*92e3bfefSLemover    XSPerfAccumulate(s"enq_state${i}", io.in.fire() && enq_state === i.U)
376*92e3bfefSLemover  }
377*92e3bfefSLemover  for (i <- 0 until (l2tlbParams.llptwsize + 1)) {
378*92e3bfefSLemover    XSPerfAccumulate(s"util${i}", PopCount(is_emptys.map(!_)) === i.U)
379*92e3bfefSLemover    XSPerfAccumulate(s"mem_util${i}", PopCount(is_mems) === i.U)
380*92e3bfefSLemover    XSPerfAccumulate(s"waiting_util${i}", PopCount(is_waiting) === i.U)
381*92e3bfefSLemover  }
382*92e3bfefSLemover  XSPerfAccumulate("mem_count", io.mem.req.fire())
383*92e3bfefSLemover  XSPerfAccumulate("mem_cycle", PopCount(is_waiting) =/= 0.U)
384*92e3bfefSLemover  XSPerfAccumulate("blocked_in", io.in.valid && !io.in.ready)
385*92e3bfefSLemover
386*92e3bfefSLemover  for (i <- 0 until l2tlbParams.llptwsize) {
387*92e3bfefSLemover    TimeOutAssert(state(i) =/= state_idle, timeOutThreshold, s"missqueue time out no out ${i}")
388*92e3bfefSLemover  }
389*92e3bfefSLemover
390*92e3bfefSLemover  val perfEvents = Seq(
391*92e3bfefSLemover    ("tlbllptw_incount           ", io.in.fire()               ),
392*92e3bfefSLemover    ("tlbllptw_inblock           ", io.in.valid && !io.in.ready),
393*92e3bfefSLemover    ("tlbllptw_memcount          ", io.mem.req.fire()          ),
394*92e3bfefSLemover    ("tlbllptw_memcycle          ", PopCount(is_waiting)       ),
395*92e3bfefSLemover  )
396*92e3bfefSLemover  generatePerfEvent()
397*92e3bfefSLemover}