109c6f1ddSLingrui98/*************************************************************************************** 209c6f1ddSLingrui98* Copyright (c) 2020-2021 Institute of Computing Technology, Chinese Academy of Sciences 309c6f1ddSLingrui98* Copyright (c) 2020-2021 Peng Cheng Laboratory 409c6f1ddSLingrui98* 509c6f1ddSLingrui98* XiangShan is licensed under Mulan PSL v2. 609c6f1ddSLingrui98* You can use this software according to the terms and conditions of the Mulan PSL v2. 709c6f1ddSLingrui98* You may obtain a copy of Mulan PSL v2 at: 809c6f1ddSLingrui98* http://license.coscl.org.cn/MulanPSL2 909c6f1ddSLingrui98* 1009c6f1ddSLingrui98* THIS SOFTWARE IS PROVIDED ON AN "AS IS" BASIS, WITHOUT WARRANTIES OF ANY KIND, 1109c6f1ddSLingrui98* EITHER EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO NON-INFRINGEMENT, 1209c6f1ddSLingrui98* MERCHANTABILITY OR FIT FOR A PARTICULAR PURPOSE. 1309c6f1ddSLingrui98* 1409c6f1ddSLingrui98* See the Mulan PSL v2 for more details. 1509c6f1ddSLingrui98***************************************************************************************/ 1609c6f1ddSLingrui98 1709c6f1ddSLingrui98package xiangshan.frontend 1809c6f1ddSLingrui98 1909c6f1ddSLingrui98import chipsalliance.rocketchip.config.Parameters 2009c6f1ddSLingrui98import chisel3._ 2109c6f1ddSLingrui98import chisel3.util._ 222a3050c2SJayimport freechips.rocketchip.rocket.RVCDecoder 2309c6f1ddSLingrui98import xiangshan._ 2409c6f1ddSLingrui98import xiangshan.cache.mmu._ 251d8f4dcbSJayimport xiangshan.frontend.icache._ 2609c6f1ddSLingrui98import utils._ 27b6982e83SLemoverimport xiangshan.backend.fu.{PMPReqBundle, PMPRespBundle} 2809c6f1ddSLingrui98 2909c6f1ddSLingrui98trait HasInstrMMIOConst extends HasXSParameter with HasIFUConst{ 3009c6f1ddSLingrui98 def mmioBusWidth = 64 3109c6f1ddSLingrui98 def mmioBusBytes = mmioBusWidth / 8 320be662e4SJay def maxInstrLen = 32 3309c6f1ddSLingrui98} 3409c6f1ddSLingrui98 3509c6f1ddSLingrui98trait HasIFUConst extends HasXSParameter{ 361d8f4dcbSJay def addrAlign(addr: UInt, bytes: Int, highest: Int): UInt = Cat(addr(highest-1, log2Ceil(bytes)), 0.U(log2Ceil(bytes).W)) 371d8f4dcbSJay def fetchQueueSize = 2 381d8f4dcbSJay 392a3050c2SJay def getBasicBlockIdx( pc: UInt, start: UInt ): UInt = { 402a3050c2SJay val byteOffset = pc - start 412a3050c2SJay (byteOffset - instBytes.U)(log2Ceil(PredictWidth),instOffsetBits) 421d8f4dcbSJay } 4309c6f1ddSLingrui98} 4409c6f1ddSLingrui98 4509c6f1ddSLingrui98class IfuToFtqIO(implicit p:Parameters) extends XSBundle { 4609c6f1ddSLingrui98 val pdWb = Valid(new PredecodeWritebackBundle) 4709c6f1ddSLingrui98} 4809c6f1ddSLingrui98 4909c6f1ddSLingrui98class FtqInterface(implicit p: Parameters) extends XSBundle { 5009c6f1ddSLingrui98 val fromFtq = Flipped(new FtqToIfuIO) 5109c6f1ddSLingrui98 val toFtq = new IfuToFtqIO 5209c6f1ddSLingrui98} 5309c6f1ddSLingrui98 540be662e4SJayclass UncacheInterface(implicit p: Parameters) extends XSBundle { 550be662e4SJay val fromUncache = Flipped(DecoupledIO(new InsUncacheResp)) 560be662e4SJay val toUncache = DecoupledIO( new InsUncacheReq ) 570be662e4SJay} 5809c6f1ddSLingrui98class NewIFUIO(implicit p: Parameters) extends XSBundle { 5909c6f1ddSLingrui98 val ftqInter = new FtqInterface 601d8f4dcbSJay val icacheInter = Vec(2, Flipped(new ICacheMainPipeBundle)) 611d8f4dcbSJay val icacheStop = Output(Bool()) 621d8f4dcbSJay val icachePerfInfo = Input(new ICachePerfInfo) 6309c6f1ddSLingrui98 val toIbuffer = Decoupled(new FetchToIBuffer) 640be662e4SJay val uncacheInter = new UncacheInterface 6572951335SLi Qianruo val frontendTrigger = Flipped(new FrontendTdataDistributeIO) 6672951335SLi Qianruo val csrTriggerEnable = Input(Vec(4, Bool())) 67a37fbf10SJay val rob_commits = Flipped(Vec(CommitWidth, Valid(new RobCommitInfo))) 68*ee175d78SJay val iTLBInter = new BlockTlbRequestIO 69*ee175d78SJay val pmp = new IPrefetchPMPBundle 7009c6f1ddSLingrui98} 7109c6f1ddSLingrui98 7209c6f1ddSLingrui98// record the situation in which fallThruAddr falls into 7309c6f1ddSLingrui98// the middle of an RVI inst 7409c6f1ddSLingrui98class LastHalfInfo(implicit p: Parameters) extends XSBundle { 7509c6f1ddSLingrui98 val valid = Bool() 7609c6f1ddSLingrui98 val middlePC = UInt(VAddrBits.W) 7709c6f1ddSLingrui98 def matchThisBlock(startAddr: UInt) = valid && middlePC === startAddr 7809c6f1ddSLingrui98} 7909c6f1ddSLingrui98 8009c6f1ddSLingrui98class IfuToPreDecode(implicit p: Parameters) extends XSBundle { 8109c6f1ddSLingrui98 val data = if(HasCExtension) Vec(PredictWidth + 1, UInt(16.W)) else Vec(PredictWidth, UInt(32.W)) 8272951335SLi Qianruo val frontendTrigger = new FrontendTdataDistributeIO 8372951335SLi Qianruo val csrTriggerEnable = Vec(4, Bool()) 842a3050c2SJay val pc = Vec(PredictWidth, UInt(VAddrBits.W)) 8509c6f1ddSLingrui98} 8609c6f1ddSLingrui98 872a3050c2SJay 882a3050c2SJayclass IfuToPredChecker(implicit p: Parameters) extends XSBundle { 892a3050c2SJay val ftqOffset = Valid(UInt(log2Ceil(PredictWidth).W)) 902a3050c2SJay val jumpOffset = Vec(PredictWidth, UInt(XLEN.W)) 912a3050c2SJay val target = UInt(VAddrBits.W) 922a3050c2SJay val instrRange = Vec(PredictWidth, Bool()) 932a3050c2SJay val instrValid = Vec(PredictWidth, Bool()) 942a3050c2SJay val pds = Vec(PredictWidth, new PreDecodeInfo) 952a3050c2SJay val pc = Vec(PredictWidth, UInt(VAddrBits.W)) 962a3050c2SJay} 972a3050c2SJay 982a3050c2SJayclass NewIFU(implicit p: Parameters) extends XSModule 992a3050c2SJay with HasICacheParameters 1002a3050c2SJay with HasIFUConst 1012a3050c2SJay with HasPdConst 1022a3050c2SJay with HasCircularQueuePtrHelper 1032a3050c2SJay with HasPerfEvents 10409c6f1ddSLingrui98{ 10509c6f1ddSLingrui98 println(s"icache ways: ${nWays} sets:${nSets}") 10609c6f1ddSLingrui98 val io = IO(new NewIFUIO) 10709c6f1ddSLingrui98 val (toFtq, fromFtq) = (io.ftqInter.toFtq, io.ftqInter.fromFtq) 1081d8f4dcbSJay val (toICache, fromICache) = (VecInit(io.icacheInter.map(_.req)), VecInit(io.icacheInter.map(_.resp))) 1090be662e4SJay val (toUncache, fromUncache) = (io.uncacheInter.toUncache , io.uncacheInter.fromUncache) 11009c6f1ddSLingrui98 11109c6f1ddSLingrui98 def isCrossLineReq(start: UInt, end: UInt): Bool = start(blockOffBits) ^ end(blockOffBits) 11209c6f1ddSLingrui98 11309c6f1ddSLingrui98 def isLastInCacheline(fallThruAddr: UInt): Bool = fallThruAddr(blockOffBits - 1, 1) === 0.U 11409c6f1ddSLingrui98 1151d8f4dcbSJay class TlbExept(implicit p: Parameters) extends XSBundle{ 1161d8f4dcbSJay val pageFault = Bool() 1171d8f4dcbSJay val accessFault = Bool() 1181d8f4dcbSJay val mmio = Bool() 119b005f7c6SJay } 12009c6f1ddSLingrui98 1212a3050c2SJay val preDecoder = Module(new PreDecode) 1222a3050c2SJay val predChecker = Module(new PredChecker) 1232a3050c2SJay val frontendTrigger = Module(new FrontendTrigger) 1242a3050c2SJay val (preDecoderIn, preDecoderOut) = (preDecoder.io.in, preDecoder.io.out) 1252a3050c2SJay val (checkerIn, checkerOut) = (predChecker.io.in, predChecker.io.out) 1261d8f4dcbSJay 127*ee175d78SJay io.iTLBInter.resp.ready := true.B 128*ee175d78SJay 12958dbdfc2SJay /** 13058dbdfc2SJay ****************************************************************************** 13158dbdfc2SJay * IFU Stage 0 13258dbdfc2SJay * - send cacheline fetch request to ICacheMainPipe 13358dbdfc2SJay ****************************************************************************** 13458dbdfc2SJay */ 13509c6f1ddSLingrui98 13609c6f1ddSLingrui98 val f0_valid = fromFtq.req.valid 13709c6f1ddSLingrui98 val f0_ftq_req = fromFtq.req.bits 13809c6f1ddSLingrui98 val f0_situation = VecInit(Seq(isCrossLineReq(f0_ftq_req.startAddr, f0_ftq_req.fallThruAddr), isLastInCacheline(f0_ftq_req.fallThruAddr))) 13909c6f1ddSLingrui98 val f0_doubleLine = f0_situation(0) || f0_situation(1) 14009c6f1ddSLingrui98 val f0_vSetIdx = VecInit(get_idx((f0_ftq_req.startAddr)), get_idx(f0_ftq_req.fallThruAddr)) 14109c6f1ddSLingrui98 val f0_fire = fromFtq.req.fire() 14209c6f1ddSLingrui98 14309c6f1ddSLingrui98 val f0_flush, f1_flush, f2_flush, f3_flush = WireInit(false.B) 14409c6f1ddSLingrui98 val from_bpu_f0_flush, from_bpu_f1_flush, from_bpu_f2_flush, from_bpu_f3_flush = WireInit(false.B) 14509c6f1ddSLingrui98 14609c6f1ddSLingrui98 from_bpu_f0_flush := fromFtq.flushFromBpu.shouldFlushByStage2(f0_ftq_req.ftqIdx) || 14709c6f1ddSLingrui98 fromFtq.flushFromBpu.shouldFlushByStage3(f0_ftq_req.ftqIdx) 14809c6f1ddSLingrui98 1492a3050c2SJay val wb_redirect , mmio_redirect, backend_redirect= WireInit(false.B) 1502a3050c2SJay val f3_wb_not_flush = WireInit(false.B) 1512a3050c2SJay 1522a3050c2SJay backend_redirect := fromFtq.redirect.valid 1532a3050c2SJay f3_flush := backend_redirect || (wb_redirect && !f3_wb_not_flush) 1542a3050c2SJay f2_flush := backend_redirect || mmio_redirect || wb_redirect 15509c6f1ddSLingrui98 f1_flush := f2_flush || from_bpu_f1_flush 15609c6f1ddSLingrui98 f0_flush := f1_flush || from_bpu_f0_flush 15709c6f1ddSLingrui98 15809c6f1ddSLingrui98 val f1_ready, f2_ready, f3_ready = WireInit(false.B) 15909c6f1ddSLingrui98 1601d8f4dcbSJay fromFtq.req.ready := toICache(0).ready && toICache(1).ready && f2_ready && GTimer() > 500.U 16109c6f1ddSLingrui98 1622a3050c2SJay toICache(0).valid := fromFtq.req.valid && !f0_flush 1631d8f4dcbSJay toICache(0).bits.vaddr := fromFtq.req.bits.startAddr 1642a3050c2SJay toICache(1).valid := fromFtq.req.valid && f0_doubleLine && !f0_flush 1651d8f4dcbSJay toICache(1).bits.vaddr := fromFtq.req.bits.fallThruAddr 16609c6f1ddSLingrui98 16758dbdfc2SJay /** <PERF> f0 fetch bubble */ 168f7c29b0aSJinYue 16958dbdfc2SJay XSPerfAccumulate("fetch_bubble_ftq_not_valid", !f0_valid ) 17058dbdfc2SJay XSPerfAccumulate("fetch_bubble_pipe_stall", f0_valid && toICache(0).ready && toICache(1).ready && !f1_ready ) 17158dbdfc2SJay XSPerfAccumulate("fetch_bubble_sram_0_busy", f0_valid && !toICache(0).ready ) 17258dbdfc2SJay XSPerfAccumulate("fetch_bubble_sram_1_busy", f0_valid && !toICache(1).ready ) 17358dbdfc2SJay 17458dbdfc2SJay 17558dbdfc2SJay /** 17658dbdfc2SJay ****************************************************************************** 17758dbdfc2SJay * IFU Stage 1 17858dbdfc2SJay * - calculate pc/half_pc/cut_ptr for every instruction 17958dbdfc2SJay ****************************************************************************** 18058dbdfc2SJay */ 1811d8f4dcbSJay 1821d8f4dcbSJay val f1_valid = RegInit(false.B) 1831d8f4dcbSJay val f1_ftq_req = RegEnable(next = f0_ftq_req, enable=f0_fire) 1841d8f4dcbSJay val f1_situation = RegEnable(next = f0_situation, enable=f0_fire) 1851d8f4dcbSJay val f1_doubleLine = RegEnable(next = f0_doubleLine, enable=f0_fire) 1861d8f4dcbSJay val f1_vSetIdx = RegEnable(next = f0_vSetIdx, enable=f0_fire) 1871d8f4dcbSJay val f1_fire = f1_valid && f1_ready 1881d8f4dcbSJay 1891d8f4dcbSJay f1_ready := f2_ready || !f1_valid 1901d8f4dcbSJay 1911d8f4dcbSJay from_bpu_f1_flush := fromFtq.flushFromBpu.shouldFlushByStage3(f1_ftq_req.ftqIdx) 1921d8f4dcbSJay 1931d8f4dcbSJay when(f1_flush) {f1_valid := false.B} 1941d8f4dcbSJay .elsewhen(f0_fire && !f0_flush) {f1_valid := true.B} 1951d8f4dcbSJay .elsewhen(f1_fire) {f1_valid := false.B} 1962a3050c2SJay 1972a3050c2SJay val f1_pc = VecInit((0 until PredictWidth).map(i => f1_ftq_req.startAddr + (i * 2).U)) 1982a3050c2SJay val f1_half_snpc = VecInit((0 until PredictWidth).map(i => f1_ftq_req.startAddr + ((i+2) * 2).U)) 1992a3050c2SJay val f1_cut_ptr = if(HasCExtension) VecInit((0 until PredictWidth + 1).map(i => Cat(0.U(1.W), f1_ftq_req.startAddr(blockOffBits-1, 1)) + i.U )) 2002a3050c2SJay else VecInit((0 until PredictWidth).map(i => Cat(0.U(1.W), f1_ftq_req.startAddr(blockOffBits-1, 2)) + i.U )) 2012a3050c2SJay 20258dbdfc2SJay /** 20358dbdfc2SJay ****************************************************************************** 20458dbdfc2SJay * IFU Stage 2 20558dbdfc2SJay * - icache response data (latched for pipeline stop) 20658dbdfc2SJay * - generate exceprion bits for every instruciton (page fault/access fault/mmio) 20758dbdfc2SJay * - generate predicted instruction range (1 means this instruciton is in this fetch packet) 20858dbdfc2SJay * - cut data from cachlines to packet instruction code 20958dbdfc2SJay * - instruction predecode and RVC expand 21058dbdfc2SJay ****************************************************************************** 21158dbdfc2SJay */ 21258dbdfc2SJay 2131d8f4dcbSJay val icacheRespAllValid = WireInit(false.B) 21409c6f1ddSLingrui98 21509c6f1ddSLingrui98 val f2_valid = RegInit(false.B) 21609c6f1ddSLingrui98 val f2_ftq_req = RegEnable(next = f1_ftq_req, enable=f1_fire) 21709c6f1ddSLingrui98 val f2_situation = RegEnable(next = f1_situation, enable=f1_fire) 21809c6f1ddSLingrui98 val f2_doubleLine = RegEnable(next = f1_doubleLine, enable=f1_fire) 2191d8f4dcbSJay val f2_vSetIdx = RegEnable(next = f1_vSetIdx, enable=f1_fire) 2201d8f4dcbSJay val f2_fire = f2_valid && f2_ready 2211d8f4dcbSJay 2221d8f4dcbSJay f2_ready := f3_ready && icacheRespAllValid || !f2_valid 2231d8f4dcbSJay //TODO: addr compare may be timing critical 2241d8f4dcbSJay val f2_icache_all_resp_wire = fromICache(0).valid && (fromICache(0).bits.vaddr === f2_ftq_req.startAddr) && ((fromICache(1).valid && (fromICache(1).bits.vaddr === f2_ftq_req.fallThruAddr)) || !f2_doubleLine) 2251d8f4dcbSJay val f2_icache_all_resp_reg = RegInit(false.B) 2261d8f4dcbSJay 2271d8f4dcbSJay icacheRespAllValid := f2_icache_all_resp_reg || f2_icache_all_resp_wire 2281d8f4dcbSJay 2291d8f4dcbSJay io.icacheStop := !f3_ready 2301d8f4dcbSJay 2311d8f4dcbSJay when(f2_flush) {f2_icache_all_resp_reg := false.B} 2321d8f4dcbSJay .elsewhen(f2_valid && f2_icache_all_resp_wire && !f3_ready) {f2_icache_all_resp_reg := true.B} 2331d8f4dcbSJay .elsewhen(f2_fire && f2_icache_all_resp_reg) {f2_icache_all_resp_reg := false.B} 23409c6f1ddSLingrui98 23509c6f1ddSLingrui98 when(f2_flush) {f2_valid := false.B} 23609c6f1ddSLingrui98 .elsewhen(f1_fire && !f1_flush) {f2_valid := true.B } 23709c6f1ddSLingrui98 .elsewhen(f2_fire) {f2_valid := false.B} 23809c6f1ddSLingrui98 2391d8f4dcbSJay val f2_cache_response_data = ResultHoldBypass(valid = f2_icache_all_resp_wire, data = VecInit(fromICache.map(_.bits.readData))) 24009c6f1ddSLingrui98 2411d8f4dcbSJay val f2_except_pf = VecInit((0 until PortNumber).map(i => fromICache(i).bits.tlbExcp.pageFault)) 2421d8f4dcbSJay val f2_except_af = VecInit((0 until PortNumber).map(i => fromICache(i).bits.tlbExcp.accessFault)) 243c0b2b8e9Srvcoresjw val f2_mmio = fromICache(0).bits.tlbExcp.mmio && !fromICache(0).bits.tlbExcp.accessFault && 244c0b2b8e9Srvcoresjw !fromICache(0).bits.tlbExcp.pageFault 2450be662e4SJay 2462a3050c2SJay val f2_pc = RegEnable(next = f1_pc, enable = f1_fire) 2472a3050c2SJay val f2_half_snpc = RegEnable(next = f1_half_snpc, enable = f1_fire) 2482a3050c2SJay val f2_cut_ptr = RegEnable(next = f1_cut_ptr, enable = f1_fire) 2492a3050c2SJay 250*ee175d78SJay val f2_resend_vaddr = RegEnable(next = f1_ftq_req.startAddr + 2.U, enable = f1_fire) 2512a3050c2SJay 2522a3050c2SJay def isNextLine(pc: UInt, startAddr: UInt) = { 2532a3050c2SJay startAddr(blockOffBits) ^ pc(blockOffBits) 2542a3050c2SJay } 2552a3050c2SJay 2562a3050c2SJay def isLastInLine(pc: UInt) = { 2572a3050c2SJay pc(blockOffBits - 1, 0) === "b111110".U 2582a3050c2SJay } 2592a3050c2SJay 2602a3050c2SJay val f2_foldpc = VecInit(f2_pc.map(i => XORFold(i(VAddrBits-1,1), MemPredPCWidth))) 2612a3050c2SJay val f2_jump_range = Fill(PredictWidth, !f2_ftq_req.ftqOffset.valid) | Fill(PredictWidth, 1.U(1.W)) >> ~f2_ftq_req.ftqOffset.bits 2622a3050c2SJay val f2_ftr_range = Fill(PredictWidth, f2_ftq_req.oversize) | Fill(PredictWidth, 1.U(1.W)) >> ~getBasicBlockIdx(f2_ftq_req.fallThruAddr, f2_ftq_req.startAddr) 2632a3050c2SJay val f2_instr_range = f2_jump_range & f2_ftr_range 2642a3050c2SJay val f2_pf_vec = VecInit((0 until PredictWidth).map(i => (!isNextLine(f2_pc(i), f2_ftq_req.startAddr) && f2_except_pf(0) || isNextLine(f2_pc(i), f2_ftq_req.startAddr) && f2_doubleLine && f2_except_pf(1)))) 2652a3050c2SJay val f2_af_vec = VecInit((0 until PredictWidth).map(i => (!isNextLine(f2_pc(i), f2_ftq_req.startAddr) && f2_except_af(0) || isNextLine(f2_pc(i), f2_ftq_req.startAddr) && f2_doubleLine && f2_except_af(1)))) 2662a3050c2SJay 2671d8f4dcbSJay val f2_paddrs = VecInit((0 until PortNumber).map(i => fromICache(i).bits.paddr)) 2681d8f4dcbSJay val f2_perf_info = io.icachePerfInfo 26909c6f1ddSLingrui98 2702a3050c2SJay def cut(cacheline: UInt, cutPtr: Vec[UInt]) : Vec[UInt] ={ 27109c6f1ddSLingrui98 if(HasCExtension){ 27209c6f1ddSLingrui98 val result = Wire(Vec(PredictWidth + 1, UInt(16.W))) 27309c6f1ddSLingrui98 val dataVec = cacheline.asTypeOf(Vec(blockBytes * 2/ 2, UInt(16.W))) 27409c6f1ddSLingrui98 (0 until PredictWidth + 1).foreach( i => 2752a3050c2SJay result(i) := dataVec(cutPtr(i)) 27609c6f1ddSLingrui98 ) 27709c6f1ddSLingrui98 result 27809c6f1ddSLingrui98 } else { 27909c6f1ddSLingrui98 val result = Wire(Vec(PredictWidth, UInt(32.W)) ) 28009c6f1ddSLingrui98 val dataVec = cacheline.asTypeOf(Vec(blockBytes * 2/ 4, UInt(32.W))) 28109c6f1ddSLingrui98 (0 until PredictWidth).foreach( i => 2822a3050c2SJay result(i) := dataVec(cutPtr(i)) 28309c6f1ddSLingrui98 ) 28409c6f1ddSLingrui98 result 28509c6f1ddSLingrui98 } 28609c6f1ddSLingrui98 } 28709c6f1ddSLingrui98 2882a3050c2SJay val f2_datas = VecInit((0 until PortNumber).map(i => f2_cache_response_data(i))) 2892a3050c2SJay val f2_cut_data = cut( Cat(f2_datas.map(cacheline => cacheline.asUInt ).reverse).asUInt, f2_cut_ptr ) 2902a3050c2SJay 29158dbdfc2SJay /** predecode (include RVC expander) */ 2922a3050c2SJay preDecoderIn.data := f2_cut_data 2932a3050c2SJay preDecoderIn.frontendTrigger := io.frontendTrigger 2942a3050c2SJay preDecoderIn.csrTriggerEnable := io.csrTriggerEnable 2952a3050c2SJay preDecoderIn.pc := f2_pc 2962a3050c2SJay 2972a3050c2SJay val f2_expd_instr = preDecoderOut.expInstr 2982a3050c2SJay val f2_pd = preDecoderOut.pd 2992a3050c2SJay val f2_jump_offset = preDecoderOut.jumpOffset 3002a3050c2SJay val f2_hasHalfValid = preDecoderOut.hasHalfValid 3012a3050c2SJay val f2_crossPageFault = VecInit((0 until PredictWidth).map(i => isLastInLine(f2_pc(i)) && !f2_except_pf(0) && f2_doubleLine && f2_except_pf(1) && !f2_pd(i).isRVC )) 3022a3050c2SJay 3031d8f4dcbSJay val predecodeOutValid = WireInit(false.B) 3041d8f4dcbSJay 30509c6f1ddSLingrui98 30658dbdfc2SJay /** 30758dbdfc2SJay ****************************************************************************** 30858dbdfc2SJay * IFU Stage 3 30958dbdfc2SJay * - handle MMIO instruciton 31058dbdfc2SJay * -send request to Uncache fetch Unit 31158dbdfc2SJay * -every packet include 1 MMIO instruction 31258dbdfc2SJay * -MMIO instructions will stop fetch pipeline until commiting from RoB 31358dbdfc2SJay * -flush to snpc (send ifu_redirect to Ftq) 31458dbdfc2SJay * - Ibuffer enqueue 31558dbdfc2SJay * - check predict result in Frontend (jalFault/retFault/notCFIFault/invalidTakenFault/targetFault) 31658dbdfc2SJay * - handle last half RVI instruction 31758dbdfc2SJay ****************************************************************************** 31858dbdfc2SJay */ 31958dbdfc2SJay 32009c6f1ddSLingrui98 val f3_valid = RegInit(false.B) 32109c6f1ddSLingrui98 val f3_ftq_req = RegEnable(next = f2_ftq_req, enable=f2_fire) 32209c6f1ddSLingrui98 val f3_situation = RegEnable(next = f2_situation, enable=f2_fire) 32309c6f1ddSLingrui98 val f3_doubleLine = RegEnable(next = f2_doubleLine, enable=f2_fire) 3241d8f4dcbSJay val f3_fire = io.toIbuffer.fire() 3251d8f4dcbSJay 3261d8f4dcbSJay f3_ready := io.toIbuffer.ready || !f3_valid 32709c6f1ddSLingrui98 32809c6f1ddSLingrui98 val f3_cut_data = RegEnable(next = f2_cut_data, enable=f2_fire) 3291d8f4dcbSJay 33009c6f1ddSLingrui98 val f3_except_pf = RegEnable(next = f2_except_pf, enable = f2_fire) 33109c6f1ddSLingrui98 val f3_except_af = RegEnable(next = f2_except_af, enable = f2_fire) 3320be662e4SJay val f3_mmio = RegEnable(next = f2_mmio , enable = f2_fire) 33309c6f1ddSLingrui98 3342a3050c2SJay val f3_expd_instr = RegEnable(next = f2_expd_instr, enable = f2_fire) 3352a3050c2SJay val f3_pd = RegEnable(next = f2_pd, enable = f2_fire) 3362a3050c2SJay val f3_jump_offset = RegEnable(next = f2_jump_offset, enable = f2_fire) 3372a3050c2SJay val f3_af_vec = RegEnable(next = f2_af_vec, enable = f2_fire) 3382a3050c2SJay val f3_pf_vec = RegEnable(next = f2_pf_vec , enable = f2_fire) 3392a3050c2SJay val f3_pc = RegEnable(next = f2_pc, enable = f2_fire) 3402a3050c2SJay val f3_half_snpc = RegEnable(next = f2_half_snpc, enable = f2_fire) 3412a3050c2SJay val f3_instr_range = RegEnable(next = f2_instr_range, enable = f2_fire) 3422a3050c2SJay val f3_foldpc = RegEnable(next = f2_foldpc, enable = f2_fire) 3432a3050c2SJay val f3_crossPageFault = RegEnable(next = f2_crossPageFault, enable = f2_fire) 3442a3050c2SJay val f3_hasHalfValid = RegEnable(next = f2_hasHalfValid, enable = f2_fire) 34509c6f1ddSLingrui98 val f3_except = VecInit((0 until 2).map{i => f3_except_pf(i) || f3_except_af(i)}) 34609c6f1ddSLingrui98 val f3_has_except = f3_valid && (f3_except_af.reduce(_||_) || f3_except_pf.reduce(_||_)) 3471d8f4dcbSJay val f3_pAddrs = RegEnable(next = f2_paddrs, enable = f2_fire) 348*ee175d78SJay val f3_resend_vaddr = RegEnable(next = f2_resend_vaddr, enable = f2_fire) 349*ee175d78SJay 350a37fbf10SJay 351a1351e5dSJay val f3_oversize_target = f3_pc.last + 2.U 352a1351e5dSJay 3532a3050c2SJay /*** MMIO State Machine***/ 354*ee175d78SJay val f3_mmio_data = Reg(Vec(2, UInt(16.W))) 355*ee175d78SJay val mmio_is_RVC = RegInit(false.B) 356*ee175d78SJay val mmio_resend_addr =RegInit(0.U(PAddrBits.W)) 357*ee175d78SJay val mmio_resend_af = RegInit(false.B) 358a37fbf10SJay 359*ee175d78SJay val m_idle :: m_sendReq :: m_waitResp :: m_sendTLB :: m_tlbResp :: m_sendPMP :: m_resendReq :: m_waitResendResp :: m_waitCommit :: m_commited :: Nil = Enum(10) 360*ee175d78SJay val mmio_state = RegInit(m_idle) 361a37fbf10SJay 3629bae7d6eSJay val f3_req_is_mmio = f3_mmio && f3_valid 3632a3050c2SJay val mmio_commit = VecInit(io.rob_commits.map{commit => commit.valid && commit.bits.ftqIdx === f3_ftq_req.ftqIdx && commit.bits.ftqOffset === 0.U}).asUInt.orR 364*ee175d78SJay val f3_mmio_req_commit = f3_req_is_mmio && mmio_state === m_commited 365a37fbf10SJay 366*ee175d78SJay val f3_mmio_to_commit = f3_req_is_mmio && mmio_state === m_waitCommit 367a37fbf10SJay val f3_mmio_to_commit_next = RegNext(f3_mmio_to_commit) 368a37fbf10SJay val f3_mmio_can_go = f3_mmio_to_commit && !f3_mmio_to_commit_next 369a37fbf10SJay 3709bae7d6eSJay val f3_ftq_flush_self = fromFtq.redirect.valid && RedirectLevel.flushItself(fromFtq.redirect.bits.level) 371167bcd01SJay val f3_ftq_flush_by_older = fromFtq.redirect.valid && isBefore(fromFtq.redirect.bits.ftqIdx, f3_ftq_req.ftqIdx) 3729bae7d6eSJay 373167bcd01SJay val f3_need_not_flush = f3_req_is_mmio && fromFtq.redirect.valid && !f3_ftq_flush_self && !f3_ftq_flush_by_older 3749bae7d6eSJay 3759bae7d6eSJay when(f3_flush && !f3_need_not_flush) {f3_valid := false.B} 376a37fbf10SJay .elsewhen(f2_fire && !f2_flush ) {f3_valid := true.B } 377a37fbf10SJay .elsewhen(io.toIbuffer.fire() && !f3_req_is_mmio) {f3_valid := false.B} 378a37fbf10SJay .elsewhen{f3_req_is_mmio && f3_mmio_req_commit} {f3_valid := false.B} 379a37fbf10SJay 380a37fbf10SJay val f3_mmio_use_seq_pc = RegInit(false.B) 381a37fbf10SJay 382a37fbf10SJay val (redirect_ftqIdx, redirect_ftqOffset) = (fromFtq.redirect.bits.ftqIdx,fromFtq.redirect.bits.ftqOffset) 383a37fbf10SJay val redirect_mmio_req = fromFtq.redirect.valid && redirect_ftqIdx === f3_ftq_req.ftqIdx && redirect_ftqOffset === 0.U 384a37fbf10SJay 385a37fbf10SJay when(RegNext(f2_fire && !f2_flush) && f3_req_is_mmio) { f3_mmio_use_seq_pc := true.B } 386a37fbf10SJay .elsewhen(redirect_mmio_req) { f3_mmio_use_seq_pc := false.B } 387a37fbf10SJay 388a37fbf10SJay f3_ready := Mux(f3_req_is_mmio, io.toIbuffer.ready && f3_mmio_req_commit || !f3_valid , io.toIbuffer.ready || !f3_valid) 389a37fbf10SJay 390*ee175d78SJay // when(fromUncache.fire()) {f3_mmio_data := fromUncache.bits.data} 391a37fbf10SJay 392a37fbf10SJay 393a37fbf10SJay switch(mmio_state){ 394*ee175d78SJay is(m_idle){ 3959bae7d6eSJay when(f3_req_is_mmio){ 396*ee175d78SJay mmio_state := m_sendReq 397a37fbf10SJay } 398a37fbf10SJay } 399a37fbf10SJay 400*ee175d78SJay is(m_sendReq){ 401*ee175d78SJay mmio_state := Mux(toUncache.fire(), m_waitResp, m_sendReq ) 402a37fbf10SJay } 403a37fbf10SJay 404*ee175d78SJay is(m_waitResp){ 405a37fbf10SJay when(fromUncache.fire()){ 406a37fbf10SJay val isRVC = fromUncache.bits.data(1,0) =/= 3.U 407*ee175d78SJay val needResend = !isRVC && f3_pAddrs(0)(2,1) === 3.U 408*ee175d78SJay mmio_state := Mux(needResend, m_sendTLB , m_waitCommit) 409*ee175d78SJay 410*ee175d78SJay mmio_is_RVC := isRVC 411*ee175d78SJay f3_mmio_data(0) := fromUncache.bits.data(15,0) 412*ee175d78SJay f3_mmio_data(1) := fromUncache.bits.data(31,16) 413a37fbf10SJay } 414a37fbf10SJay } 415a37fbf10SJay 416*ee175d78SJay is(m_sendTLB){ 417*ee175d78SJay mmio_state := m_tlbResp 418a37fbf10SJay } 419a37fbf10SJay 420*ee175d78SJay is(m_tlbResp){ 421*ee175d78SJay mmio_state := m_sendPMP 422*ee175d78SJay mmio_resend_addr := io.iTLBInter.resp.bits.paddr 423*ee175d78SJay } 424*ee175d78SJay 425*ee175d78SJay is(m_sendPMP){ 426*ee175d78SJay val pmpExcpAF = io.pmp.resp.instr 427*ee175d78SJay mmio_state := Mux(pmpExcpAF, m_waitCommit , m_resendReq) 428*ee175d78SJay mmio_resend_af := pmpExcpAF 429*ee175d78SJay } 430*ee175d78SJay 431*ee175d78SJay is(m_resendReq){ 432*ee175d78SJay mmio_state := Mux(toUncache.fire(), m_waitResendResp, m_resendReq ) 433*ee175d78SJay } 434*ee175d78SJay 435*ee175d78SJay is(m_waitResendResp){ 436a37fbf10SJay when(fromUncache.fire()){ 437*ee175d78SJay mmio_state := m_waitCommit 438*ee175d78SJay f3_mmio_data(1) := fromUncache.bits.data(15,0) 439a37fbf10SJay } 440a37fbf10SJay } 441a37fbf10SJay 442*ee175d78SJay is(m_waitCommit){ 4432a3050c2SJay when(mmio_commit){ 444*ee175d78SJay mmio_state := m_commited 445a37fbf10SJay } 446a37fbf10SJay } 4472a3050c2SJay 448*ee175d78SJay //normal mmio instruction 449*ee175d78SJay is(m_commited){ 450*ee175d78SJay mmio_state := m_idle 451*ee175d78SJay mmio_is_RVC := false.B 452*ee175d78SJay mmio_resend_addr := 0.U 4532a3050c2SJay } 454a37fbf10SJay } 455a37fbf10SJay 456*ee175d78SJay //exception or flush by older branch prediction 457167bcd01SJay when(f3_ftq_flush_self || f3_ftq_flush_by_older) { 458*ee175d78SJay mmio_state := m_idle 459*ee175d78SJay mmio_is_RVC := false.B 460*ee175d78SJay mmio_resend_addr := 0.U 461*ee175d78SJay mmio_resend_af := false.B 462*ee175d78SJay f3_mmio_data.map(_ := 0.U) 4639bae7d6eSJay } 4649bae7d6eSJay 465*ee175d78SJay toUncache.valid := ((mmio_state === m_sendReq) || (mmio_state === m_resendReq)) && f3_req_is_mmio 466*ee175d78SJay toUncache.bits.addr := Mux((mmio_state === m_resendReq), mmio_resend_addr, f3_pAddrs(0)) 467a37fbf10SJay fromUncache.ready := true.B 468a37fbf10SJay 469*ee175d78SJay io.iTLBInter.req.valid := (mmio_state === m_sendTLB) && f3_req_is_mmio 470*ee175d78SJay io.iTLBInter.req.bits.size := 3.U 471*ee175d78SJay io.iTLBInter.req.bits.vaddr := f3_resend_vaddr 472*ee175d78SJay io.iTLBInter.req.bits.debug.pc := f3_resend_vaddr 473*ee175d78SJay 474*ee175d78SJay io.iTLBInter.req.bits.cmd := TlbCmd.exec 475*ee175d78SJay io.iTLBInter.req.bits.robIdx := DontCare 476*ee175d78SJay io.iTLBInter.req.bits.debug.isFirstIssue := DontCare 477*ee175d78SJay 478*ee175d78SJay io.pmp.req.valid := (mmio_state === m_sendPMP) && f3_req_is_mmio 479*ee175d78SJay io.pmp.req.bits.addr := mmio_resend_addr 480*ee175d78SJay io.pmp.req.bits.size := 3.U 481*ee175d78SJay io.pmp.req.bits.cmd := TlbCmd.exec 48209c6f1ddSLingrui98 4832a3050c2SJay val f3_lastHalf = RegInit(0.U.asTypeOf(new LastHalfInfo)) 48409c6f1ddSLingrui98 48509c6f1ddSLingrui98 val f3_predecode_range = VecInit(preDecoderOut.pd.map(inst => inst.valid)).asUInt 4860be662e4SJay val f3_mmio_range = VecInit((0 until PredictWidth).map(i => if(i ==0) true.B else false.B)) 4872a3050c2SJay val f3_instr_valid = Wire(Vec(PredictWidth, Bool())) 48809c6f1ddSLingrui98 4892a3050c2SJay /*** prediction result check ***/ 4902a3050c2SJay checkerIn.ftqOffset := f3_ftq_req.ftqOffset 4912a3050c2SJay checkerIn.jumpOffset := f3_jump_offset 4922a3050c2SJay checkerIn.target := f3_ftq_req.target 4932a3050c2SJay checkerIn.instrRange := f3_instr_range.asTypeOf(Vec(PredictWidth, Bool())) 4942a3050c2SJay checkerIn.instrValid := f3_instr_valid.asTypeOf(Vec(PredictWidth, Bool())) 4952a3050c2SJay checkerIn.pds := f3_pd 4962a3050c2SJay checkerIn.pc := f3_pc 4972a3050c2SJay 49858dbdfc2SJay /*** handle half RVI in the last 2 Bytes ***/ 4992a3050c2SJay 5002a3050c2SJay def hasLastHalf(idx: UInt) = { 5012a3050c2SJay !f3_pd(idx).isRVC && checkerOut.fixedRange(idx) && f3_instr_valid(idx) && !checkerOut.fixedTaken(idx) && !checkerOut.fixedMissPred(idx) && ! f3_req_is_mmio && !f3_ftq_req.oversize 5022a3050c2SJay } 5032a3050c2SJay 5042a3050c2SJay val f3_last_validIdx = ~ParallelPriorityEncoder(checkerOut.fixedRange.reverse) 5052a3050c2SJay 5062a3050c2SJay val f3_hasLastHalf = hasLastHalf((PredictWidth - 1).U) 5072a3050c2SJay val f3_false_lastHalf = hasLastHalf(f3_last_validIdx) 5082a3050c2SJay val f3_false_snpc = f3_half_snpc(f3_last_validIdx) 5092a3050c2SJay 5102a3050c2SJay val f3_lastHalf_mask = VecInit((0 until PredictWidth).map( i => if(i ==0) false.B else true.B )).asUInt() 5112a3050c2SJay 5122a3050c2SJay when (f3_flush) { 5132a3050c2SJay f3_lastHalf.valid := false.B 5142a3050c2SJay }.elsewhen (f3_fire) { 5152a3050c2SJay f3_lastHalf.valid := f3_hasLastHalf 5162a3050c2SJay f3_lastHalf.middlePC := f3_ftq_req.fallThruAddr 5172a3050c2SJay } 5182a3050c2SJay 5192a3050c2SJay f3_instr_valid := Mux(f3_lastHalf.valid,f3_hasHalfValid ,VecInit(f3_pd.map(inst => inst.valid))) 5202a3050c2SJay 5212a3050c2SJay /*** frontend Trigger ***/ 5222a3050c2SJay frontendTrigger.io.pds := f3_pd 5232a3050c2SJay frontendTrigger.io.pc := f3_pc 5242a3050c2SJay frontendTrigger.io.data := f3_cut_data 5252a3050c2SJay 5262a3050c2SJay frontendTrigger.io.frontendTrigger := io.frontendTrigger 5272a3050c2SJay frontendTrigger.io.csrTriggerEnable := io.csrTriggerEnable 5282a3050c2SJay 5292a3050c2SJay val f3_triggered = frontendTrigger.io.triggered 5302a3050c2SJay 5312a3050c2SJay /*** send to Ibuffer ***/ 5322a3050c2SJay 5332a3050c2SJay io.toIbuffer.valid := f3_valid && (!f3_req_is_mmio || f3_mmio_can_go) && !f3_flush 5342a3050c2SJay io.toIbuffer.bits.instrs := f3_expd_instr 5352a3050c2SJay io.toIbuffer.bits.valid := f3_instr_valid.asUInt 5362a3050c2SJay io.toIbuffer.bits.enqEnable := checkerOut.fixedRange.asUInt & f3_instr_valid.asUInt 5372a3050c2SJay io.toIbuffer.bits.pd := f3_pd 53809c6f1ddSLingrui98 io.toIbuffer.bits.ftqPtr := f3_ftq_req.ftqIdx 5392a3050c2SJay io.toIbuffer.bits.pc := f3_pc 5402a3050c2SJay io.toIbuffer.bits.ftqOffset.zipWithIndex.map{case(a, i) => a.bits := i.U; a.valid := checkerOut.fixedTaken(i) && !f3_req_is_mmio} 5412a3050c2SJay io.toIbuffer.bits.foldpc := f3_foldpc 5422a3050c2SJay io.toIbuffer.bits.ipf := f3_pf_vec 5432a3050c2SJay io.toIbuffer.bits.acf := f3_af_vec 5442a3050c2SJay io.toIbuffer.bits.crossPageIPFFix := f3_crossPageFault 5452a3050c2SJay io.toIbuffer.bits.triggered := f3_triggered 5462a3050c2SJay 5472a3050c2SJay val lastHalfMask = VecInit((0 until PredictWidth).map(i => if(i ==0) false.B else true.B)) 5482a3050c2SJay when(f3_lastHalf.valid){ 5492a3050c2SJay io.toIbuffer.bits.enqEnable := checkerOut.fixedRange.asUInt & f3_instr_valid.asUInt & lastHalfMask.asUInt 5502a3050c2SJay io.toIbuffer.bits.valid := f3_lastHalf_mask & f3_instr_valid.asUInt 5512a3050c2SJay } 5522a3050c2SJay 5532a3050c2SJay /** external predecode for MMIO instruction */ 5542a3050c2SJay when(f3_req_is_mmio){ 555*ee175d78SJay val inst = Cat(f3_mmio_data(1), f3_mmio_data(0)) 5562a3050c2SJay val currentIsRVC = isRVC(inst) 5572a3050c2SJay 5582a3050c2SJay val brType::isCall::isRet::Nil = brInfo(inst) 5592a3050c2SJay val jalOffset = jal_offset(inst, currentIsRVC) 5602a3050c2SJay val brOffset = br_offset(inst, currentIsRVC) 5612a3050c2SJay 5622a3050c2SJay io.toIbuffer.bits.instrs (0) := new RVCDecoder(inst, XLEN).decode.bits 5632a3050c2SJay 5642a3050c2SJay io.toIbuffer.bits.pd(0).valid := true.B 5652a3050c2SJay io.toIbuffer.bits.pd(0).isRVC := currentIsRVC 5662a3050c2SJay io.toIbuffer.bits.pd(0).brType := brType 5672a3050c2SJay io.toIbuffer.bits.pd(0).isCall := isCall 5682a3050c2SJay io.toIbuffer.bits.pd(0).isRet := isRet 5692a3050c2SJay 570*ee175d78SJay io.toIbuffer.bits.acf(0) := mmio_resend_af 571*ee175d78SJay 5722a3050c2SJay io.toIbuffer.bits.enqEnable := f3_mmio_range.asUInt 5732a3050c2SJay } 5742a3050c2SJay 57509c6f1ddSLingrui98 57609c6f1ddSLingrui98 //Write back to Ftq 577a37fbf10SJay val f3_cache_fetch = f3_valid && !(f2_fire && !f2_flush) 578a37fbf10SJay val finishFetchMaskReg = RegNext(f3_cache_fetch) 579a37fbf10SJay 5802a3050c2SJay val mmioFlushWb = Wire(Valid(new PredecodeWritebackBundle)) 5810be662e4SJay val f3_mmio_missOffset = Wire(ValidUndirectioned(UInt(log2Ceil(PredictWidth).W))) 582a37fbf10SJay f3_mmio_missOffset.valid := f3_req_is_mmio 5830be662e4SJay f3_mmio_missOffset.bits := 0.U 5840be662e4SJay 585*ee175d78SJay mmioFlushWb.valid := (f3_req_is_mmio && mmio_state === m_waitCommit && RegNext(fromUncache.fire()) && f3_mmio_use_seq_pc) 5862a3050c2SJay mmioFlushWb.bits.pc := f3_pc 5872a3050c2SJay mmioFlushWb.bits.pd := f3_pd 5882a3050c2SJay mmioFlushWb.bits.pd.zipWithIndex.map{case(instr,i) => instr.valid := f3_mmio_range(i)} 5892a3050c2SJay mmioFlushWb.bits.ftqIdx := f3_ftq_req.ftqIdx 5902a3050c2SJay mmioFlushWb.bits.ftqOffset := f3_ftq_req.ftqOffset.bits 5912a3050c2SJay mmioFlushWb.bits.misOffset := f3_mmio_missOffset 5922a3050c2SJay mmioFlushWb.bits.cfiOffset := DontCare 593*ee175d78SJay mmioFlushWb.bits.target := Mux(mmio_is_RVC, f3_ftq_req.startAddr + 2.U , f3_ftq_req.startAddr + 4.U) 5942a3050c2SJay mmioFlushWb.bits.jalTarget := DontCare 5952a3050c2SJay mmioFlushWb.bits.instrRange := f3_mmio_range 59609c6f1ddSLingrui98 597*ee175d78SJay mmio_redirect := (f3_req_is_mmio && mmio_state === m_waitCommit && RegNext(fromUncache.fire()) && f3_mmio_use_seq_pc) 5982a3050c2SJay 59958dbdfc2SJay /** 60058dbdfc2SJay ****************************************************************************** 60158dbdfc2SJay * IFU Write Back Stage 60258dbdfc2SJay * - write back predecode information to Ftq to update 60358dbdfc2SJay * - redirect if found fault prediction 60458dbdfc2SJay * - redirect if has false hit last half (last PC is not start + 32 Bytes, but in the midle of an notCFI RVI instruction) 60558dbdfc2SJay ****************************************************************************** 6062a3050c2SJay */ 60758dbdfc2SJay 6082a3050c2SJay val wb_valid = RegNext(RegNext(f2_fire && !f2_flush) && !f3_req_is_mmio && !f3_flush) 6092a3050c2SJay val wb_ftq_req = RegNext(f3_ftq_req) 6102a3050c2SJay 6112a3050c2SJay val wb_check_result = RegNext(checkerOut) 6122a3050c2SJay val wb_instr_range = RegNext(io.toIbuffer.bits.enqEnable) 6132a3050c2SJay val wb_pc = RegNext(f3_pc) 6142a3050c2SJay val wb_pd = RegNext(f3_pd) 6152a3050c2SJay val wb_instr_valid = RegNext(f3_instr_valid) 6162a3050c2SJay 6172a3050c2SJay /* false hit lastHalf */ 6182a3050c2SJay val wb_lastIdx = RegNext(f3_last_validIdx) 6192a3050c2SJay val wb_false_lastHalf = RegNext(f3_false_lastHalf) && wb_lastIdx =/= (PredictWidth - 1).U 6202a3050c2SJay val wb_false_target = RegNext(f3_false_snpc) 6212a3050c2SJay 6222a3050c2SJay val wb_half_flush = wb_false_lastHalf 6232a3050c2SJay val wb_half_target = wb_false_target 6242a3050c2SJay 625a1351e5dSJay /* false oversize */ 626a1351e5dSJay val lastIsRVC = wb_instr_range.asTypeOf(Vec(PredictWidth,Bool())).last && wb_pd.last.isRVC 627a1351e5dSJay val lastIsRVI = wb_instr_range.asTypeOf(Vec(PredictWidth,Bool()))(PredictWidth - 2) && !wb_pd(PredictWidth - 2).isRVC 628a1351e5dSJay val lastTaken = wb_check_result.fixedTaken.last 629a1351e5dSJay val wb_false_oversize = wb_valid && wb_ftq_req.oversize && (lastIsRVC || lastIsRVI) && !lastTaken 630a1351e5dSJay val wb_oversize_target = RegNext(f3_oversize_target) 631a1351e5dSJay 632a1351e5dSJay when(wb_valid){ 633a1351e5dSJay assert(!wb_false_oversize || !wb_half_flush, "False oversize and false half should be exclusive. ") 634a1351e5dSJay } 635a1351e5dSJay 6362a3050c2SJay f3_wb_not_flush := wb_ftq_req.ftqIdx === f3_ftq_req.ftqIdx && f3_valid && wb_valid 6372a3050c2SJay 6382a3050c2SJay val checkFlushWb = Wire(Valid(new PredecodeWritebackBundle)) 6392a3050c2SJay checkFlushWb.valid := wb_valid 6402a3050c2SJay checkFlushWb.bits.pc := wb_pc 6412a3050c2SJay checkFlushWb.bits.pd := wb_pd 6422a3050c2SJay checkFlushWb.bits.pd.zipWithIndex.map{case(instr,i) => instr.valid := wb_instr_valid(i)} 6432a3050c2SJay checkFlushWb.bits.ftqIdx := wb_ftq_req.ftqIdx 6442a3050c2SJay checkFlushWb.bits.ftqOffset := wb_ftq_req.ftqOffset.bits 645a1351e5dSJay checkFlushWb.bits.misOffset.valid := ParallelOR(wb_check_result.fixedMissPred) || wb_half_flush || wb_false_oversize 6462a3050c2SJay checkFlushWb.bits.misOffset.bits := Mux(wb_half_flush, (PredictWidth - 1).U, ParallelPriorityEncoder(wb_check_result.fixedMissPred)) 6472a3050c2SJay checkFlushWb.bits.cfiOffset.valid := ParallelOR(wb_check_result.fixedTaken) 6482a3050c2SJay checkFlushWb.bits.cfiOffset.bits := ParallelPriorityEncoder(wb_check_result.fixedTaken) 649a1351e5dSJay checkFlushWb.bits.target := Mux(wb_false_oversize, wb_oversize_target, 650a1351e5dSJay Mux(wb_half_flush, wb_half_target, wb_check_result.fixedTarget(ParallelPriorityEncoder(wb_check_result.fixedMissPred)))) 6512a3050c2SJay checkFlushWb.bits.jalTarget := wb_check_result.fixedTarget(ParallelPriorityEncoder(VecInit(wb_pd.map{pd => pd.isJal }))) 6522a3050c2SJay checkFlushWb.bits.instrRange := wb_instr_range.asTypeOf(Vec(PredictWidth, Bool())) 6532a3050c2SJay 6542a3050c2SJay toFtq.pdWb := Mux(f3_req_is_mmio, mmioFlushWb, checkFlushWb) 6552a3050c2SJay 6562a3050c2SJay wb_redirect := checkFlushWb.bits.misOffset.valid && wb_valid 65709c6f1ddSLingrui98 6581d8f4dcbSJay 6591d8f4dcbSJay /** performance counter */ 6601d8f4dcbSJay val f3_perf_info = RegEnable(next = f2_perf_info, enable = f2_fire) 6611d8f4dcbSJay val f3_req_0 = io.toIbuffer.fire() 6621d8f4dcbSJay val f3_req_1 = io.toIbuffer.fire() && f3_doubleLine 6631d8f4dcbSJay val f3_hit_0 = io.toIbuffer.fire() && f3_perf_info.bank_hit(0) 6641d8f4dcbSJay val f3_hit_1 = io.toIbuffer.fire() && f3_doubleLine & f3_perf_info.bank_hit(1) 6651d8f4dcbSJay val f3_hit = f3_perf_info.hit 666cd365d4cSrvcoresjw val perfEvents = Seq( 6672a3050c2SJay ("frontendFlush ", wb_redirect ), 668cd365d4cSrvcoresjw ("ifu_req ", io.toIbuffer.fire() ), 6691d8f4dcbSJay ("ifu_miss ", io.toIbuffer.fire() && !f3_perf_info.hit ), 670cd365d4cSrvcoresjw ("ifu_req_cacheline_0 ", f3_req_0 ), 671cd365d4cSrvcoresjw ("ifu_req_cacheline_1 ", f3_req_1 ), 672cd365d4cSrvcoresjw ("ifu_req_cacheline_0_hit ", f3_hit_1 ), 673cd365d4cSrvcoresjw ("ifu_req_cacheline_1_hit ", f3_hit_1 ), 6741d8f4dcbSJay ("only_0_hit ", f3_perf_info.only_0_hit && io.toIbuffer.fire() ), 6751d8f4dcbSJay ("only_0_miss ", f3_perf_info.only_0_miss && io.toIbuffer.fire() ), 6761d8f4dcbSJay ("hit_0_hit_1 ", f3_perf_info.hit_0_hit_1 && io.toIbuffer.fire() ), 6771d8f4dcbSJay ("hit_0_miss_1 ", f3_perf_info.hit_0_miss_1 && io.toIbuffer.fire() ), 6781d8f4dcbSJay ("miss_0_hit_1 ", f3_perf_info.miss_0_hit_1 && io.toIbuffer.fire() ), 6791d8f4dcbSJay ("miss_0_miss_1 ", f3_perf_info.miss_0_miss_1 && io.toIbuffer.fire() ), 680cd365d4cSrvcoresjw ("cross_line_block ", io.toIbuffer.fire() && f3_situation(0) ), 681cd365d4cSrvcoresjw ("fall_through_is_cacheline_end", io.toIbuffer.fire() && f3_situation(1) ), 682cd365d4cSrvcoresjw ) 6831ca0e4f3SYinan Xu generatePerfEvent() 68409c6f1ddSLingrui98 685f7c29b0aSJinYue XSPerfAccumulate("ifu_req", io.toIbuffer.fire() ) 686f7c29b0aSJinYue XSPerfAccumulate("ifu_miss", io.toIbuffer.fire() && !f3_hit ) 687f7c29b0aSJinYue XSPerfAccumulate("ifu_req_cacheline_0", f3_req_0 ) 688f7c29b0aSJinYue XSPerfAccumulate("ifu_req_cacheline_1", f3_req_1 ) 689f7c29b0aSJinYue XSPerfAccumulate("ifu_req_cacheline_0_hit", f3_hit_0 ) 690f7c29b0aSJinYue XSPerfAccumulate("ifu_req_cacheline_1_hit", f3_hit_1 ) 6912a3050c2SJay XSPerfAccumulate("frontendFlush", wb_redirect ) 6921d8f4dcbSJay XSPerfAccumulate("only_0_hit", f3_perf_info.only_0_hit && io.toIbuffer.fire() ) 6931d8f4dcbSJay XSPerfAccumulate("only_0_miss", f3_perf_info.only_0_miss && io.toIbuffer.fire() ) 6941d8f4dcbSJay XSPerfAccumulate("hit_0_hit_1", f3_perf_info.hit_0_hit_1 && io.toIbuffer.fire() ) 6951d8f4dcbSJay XSPerfAccumulate("hit_0_miss_1", f3_perf_info.hit_0_miss_1 && io.toIbuffer.fire() ) 6961d8f4dcbSJay XSPerfAccumulate("miss_0_hit_1", f3_perf_info.miss_0_hit_1 && io.toIbuffer.fire() ) 6971d8f4dcbSJay XSPerfAccumulate("miss_0_miss_1", f3_perf_info.miss_0_miss_1 && io.toIbuffer.fire() ) 698f7c29b0aSJinYue XSPerfAccumulate("cross_line_block", io.toIbuffer.fire() && f3_situation(0) ) 699f7c29b0aSJinYue XSPerfAccumulate("fall_through_is_cacheline_end", io.toIbuffer.fire() && f3_situation(1) ) 70009c6f1ddSLingrui98} 701