109c6f1ddSLingrui98/*************************************************************************************** 209c6f1ddSLingrui98* Copyright (c) 2020-2021 Institute of Computing Technology, Chinese Academy of Sciences 309c6f1ddSLingrui98* Copyright (c) 2020-2021 Peng Cheng Laboratory 409c6f1ddSLingrui98* 509c6f1ddSLingrui98* XiangShan is licensed under Mulan PSL v2. 609c6f1ddSLingrui98* You can use this software according to the terms and conditions of the Mulan PSL v2. 709c6f1ddSLingrui98* You may obtain a copy of Mulan PSL v2 at: 809c6f1ddSLingrui98* http://license.coscl.org.cn/MulanPSL2 909c6f1ddSLingrui98* 1009c6f1ddSLingrui98* THIS SOFTWARE IS PROVIDED ON AN "AS IS" BASIS, WITHOUT WARRANTIES OF ANY KIND, 1109c6f1ddSLingrui98* EITHER EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO NON-INFRINGEMENT, 1209c6f1ddSLingrui98* MERCHANTABILITY OR FIT FOR A PARTICULAR PURPOSE. 1309c6f1ddSLingrui98* 1409c6f1ddSLingrui98* See the Mulan PSL v2 for more details. 1509c6f1ddSLingrui98***************************************************************************************/ 1609c6f1ddSLingrui98 1709c6f1ddSLingrui98package xiangshan.frontend 1809c6f1ddSLingrui98 1909c6f1ddSLingrui98import chipsalliance.rocketchip.config.Parameters 2009c6f1ddSLingrui98import chisel3._ 2109c6f1ddSLingrui98import chisel3.util._ 222a3050c2SJayimport freechips.rocketchip.rocket.RVCDecoder 2309c6f1ddSLingrui98import xiangshan._ 2409c6f1ddSLingrui98import xiangshan.cache.mmu._ 251d8f4dcbSJayimport xiangshan.frontend.icache._ 2609c6f1ddSLingrui98import utils._ 27b6982e83SLemoverimport xiangshan.backend.fu.{PMPReqBundle, PMPRespBundle} 2809c6f1ddSLingrui98 2909c6f1ddSLingrui98trait HasInstrMMIOConst extends HasXSParameter with HasIFUConst{ 3009c6f1ddSLingrui98 def mmioBusWidth = 64 3109c6f1ddSLingrui98 def mmioBusBytes = mmioBusWidth / 8 320be662e4SJay def maxInstrLen = 32 3309c6f1ddSLingrui98} 3409c6f1ddSLingrui98 3509c6f1ddSLingrui98trait HasIFUConst extends HasXSParameter{ 361d8f4dcbSJay def addrAlign(addr: UInt, bytes: Int, highest: Int): UInt = Cat(addr(highest-1, log2Ceil(bytes)), 0.U(log2Ceil(bytes).W)) 371d8f4dcbSJay def fetchQueueSize = 2 381d8f4dcbSJay 392a3050c2SJay def getBasicBlockIdx( pc: UInt, start: UInt ): UInt = { 402a3050c2SJay val byteOffset = pc - start 412a3050c2SJay (byteOffset - instBytes.U)(log2Ceil(PredictWidth),instOffsetBits) 421d8f4dcbSJay } 4309c6f1ddSLingrui98} 4409c6f1ddSLingrui98 4509c6f1ddSLingrui98class IfuToFtqIO(implicit p:Parameters) extends XSBundle { 4609c6f1ddSLingrui98 val pdWb = Valid(new PredecodeWritebackBundle) 4709c6f1ddSLingrui98} 4809c6f1ddSLingrui98 4909c6f1ddSLingrui98class FtqInterface(implicit p: Parameters) extends XSBundle { 5009c6f1ddSLingrui98 val fromFtq = Flipped(new FtqToIfuIO) 5109c6f1ddSLingrui98 val toFtq = new IfuToFtqIO 5209c6f1ddSLingrui98} 5309c6f1ddSLingrui98 540be662e4SJayclass UncacheInterface(implicit p: Parameters) extends XSBundle { 550be662e4SJay val fromUncache = Flipped(DecoupledIO(new InsUncacheResp)) 560be662e4SJay val toUncache = DecoupledIO( new InsUncacheReq ) 570be662e4SJay} 5809c6f1ddSLingrui98class NewIFUIO(implicit p: Parameters) extends XSBundle { 5909c6f1ddSLingrui98 val ftqInter = new FtqInterface 601d8f4dcbSJay val icacheInter = Vec(2, Flipped(new ICacheMainPipeBundle)) 611d8f4dcbSJay val icacheStop = Output(Bool()) 621d8f4dcbSJay val icachePerfInfo = Input(new ICachePerfInfo) 6309c6f1ddSLingrui98 val toIbuffer = Decoupled(new FetchToIBuffer) 640be662e4SJay val uncacheInter = new UncacheInterface 6572951335SLi Qianruo val frontendTrigger = Flipped(new FrontendTdataDistributeIO) 6672951335SLi Qianruo val csrTriggerEnable = Input(Vec(4, Bool())) 67a37fbf10SJay val rob_commits = Flipped(Vec(CommitWidth, Valid(new RobCommitInfo))) 68ee175d78SJay val iTLBInter = new BlockTlbRequestIO 6956788a33SJinYue val pmp = new ICachePMPBundle 7009c6f1ddSLingrui98} 7109c6f1ddSLingrui98 7209c6f1ddSLingrui98// record the situation in which fallThruAddr falls into 7309c6f1ddSLingrui98// the middle of an RVI inst 7409c6f1ddSLingrui98class LastHalfInfo(implicit p: Parameters) extends XSBundle { 7509c6f1ddSLingrui98 val valid = Bool() 7609c6f1ddSLingrui98 val middlePC = UInt(VAddrBits.W) 7709c6f1ddSLingrui98 def matchThisBlock(startAddr: UInt) = valid && middlePC === startAddr 7809c6f1ddSLingrui98} 7909c6f1ddSLingrui98 8009c6f1ddSLingrui98class IfuToPreDecode(implicit p: Parameters) extends XSBundle { 8109c6f1ddSLingrui98 val data = if(HasCExtension) Vec(PredictWidth + 1, UInt(16.W)) else Vec(PredictWidth, UInt(32.W)) 8272951335SLi Qianruo val frontendTrigger = new FrontendTdataDistributeIO 8372951335SLi Qianruo val csrTriggerEnable = Vec(4, Bool()) 842a3050c2SJay val pc = Vec(PredictWidth, UInt(VAddrBits.W)) 8509c6f1ddSLingrui98} 8609c6f1ddSLingrui98 872a3050c2SJay 882a3050c2SJayclass IfuToPredChecker(implicit p: Parameters) extends XSBundle { 892a3050c2SJay val ftqOffset = Valid(UInt(log2Ceil(PredictWidth).W)) 902a3050c2SJay val jumpOffset = Vec(PredictWidth, UInt(XLEN.W)) 912a3050c2SJay val target = UInt(VAddrBits.W) 922a3050c2SJay val instrRange = Vec(PredictWidth, Bool()) 932a3050c2SJay val instrValid = Vec(PredictWidth, Bool()) 942a3050c2SJay val pds = Vec(PredictWidth, new PreDecodeInfo) 952a3050c2SJay val pc = Vec(PredictWidth, UInt(VAddrBits.W)) 962a3050c2SJay} 972a3050c2SJay 982a3050c2SJayclass NewIFU(implicit p: Parameters) extends XSModule 992a3050c2SJay with HasICacheParameters 1002a3050c2SJay with HasIFUConst 1012a3050c2SJay with HasPdConst 102167bcd01SJay with HasCircularQueuePtrHelper 1032a3050c2SJay with HasPerfEvents 10409c6f1ddSLingrui98{ 10509c6f1ddSLingrui98 val io = IO(new NewIFUIO) 10609c6f1ddSLingrui98 val (toFtq, fromFtq) = (io.ftqInter.toFtq, io.ftqInter.fromFtq) 1071d8f4dcbSJay val (toICache, fromICache) = (VecInit(io.icacheInter.map(_.req)), VecInit(io.icacheInter.map(_.resp))) 1080be662e4SJay val (toUncache, fromUncache) = (io.uncacheInter.toUncache , io.uncacheInter.fromUncache) 10909c6f1ddSLingrui98 11009c6f1ddSLingrui98 def isCrossLineReq(start: UInt, end: UInt): Bool = start(blockOffBits) ^ end(blockOffBits) 11109c6f1ddSLingrui98 11234a88126SJinYue def isLastInCacheline(addr: UInt): Bool = addr(blockOffBits - 1, 1) === 0.U 11309c6f1ddSLingrui98 1141d8f4dcbSJay class TlbExept(implicit p: Parameters) extends XSBundle{ 1151d8f4dcbSJay val pageFault = Bool() 1161d8f4dcbSJay val accessFault = Bool() 1171d8f4dcbSJay val mmio = Bool() 118b005f7c6SJay } 11909c6f1ddSLingrui98 1202a3050c2SJay val preDecoder = Module(new PreDecode) 1212a3050c2SJay val predChecker = Module(new PredChecker) 1222a3050c2SJay val frontendTrigger = Module(new FrontendTrigger) 1232a3050c2SJay val (preDecoderIn, preDecoderOut) = (preDecoder.io.in, preDecoder.io.out) 1242a3050c2SJay val (checkerIn, checkerOut) = (predChecker.io.in, predChecker.io.out) 1251d8f4dcbSJay 126ee175d78SJay io.iTLBInter.resp.ready := true.B 127ee175d78SJay 12858dbdfc2SJay /** 12958dbdfc2SJay ****************************************************************************** 13058dbdfc2SJay * IFU Stage 0 13158dbdfc2SJay * - send cacheline fetch request to ICacheMainPipe 13258dbdfc2SJay ****************************************************************************** 13358dbdfc2SJay */ 13409c6f1ddSLingrui98 13509c6f1ddSLingrui98 val f0_valid = fromFtq.req.valid 13609c6f1ddSLingrui98 val f0_ftq_req = fromFtq.req.bits 1376ce52296SJinYue val f0_doubleLine = fromFtq.req.bits.crossCacheline 13834a88126SJinYue val f0_vSetIdx = VecInit(get_idx((f0_ftq_req.startAddr)), get_idx(f0_ftq_req.nextlineStart)) 13909c6f1ddSLingrui98 val f0_fire = fromFtq.req.fire() 14009c6f1ddSLingrui98 14109c6f1ddSLingrui98 val f0_flush, f1_flush, f2_flush, f3_flush = WireInit(false.B) 14209c6f1ddSLingrui98 val from_bpu_f0_flush, from_bpu_f1_flush, from_bpu_f2_flush, from_bpu_f3_flush = WireInit(false.B) 14309c6f1ddSLingrui98 144cb4f77ceSLingrui98 from_bpu_f0_flush := fromFtq.flushFromBpu.shouldFlushByStage2(f0_ftq_req.ftqIdx) || 145cb4f77ceSLingrui98 fromFtq.flushFromBpu.shouldFlushByStage3(f0_ftq_req.ftqIdx) 14609c6f1ddSLingrui98 1472a3050c2SJay val wb_redirect , mmio_redirect, backend_redirect= WireInit(false.B) 1482a3050c2SJay val f3_wb_not_flush = WireInit(false.B) 1492a3050c2SJay 1502a3050c2SJay backend_redirect := fromFtq.redirect.valid 1512a3050c2SJay f3_flush := backend_redirect || (wb_redirect && !f3_wb_not_flush) 1522a3050c2SJay f2_flush := backend_redirect || mmio_redirect || wb_redirect 15309c6f1ddSLingrui98 f1_flush := f2_flush || from_bpu_f1_flush 15409c6f1ddSLingrui98 f0_flush := f1_flush || from_bpu_f0_flush 15509c6f1ddSLingrui98 15609c6f1ddSLingrui98 val f1_ready, f2_ready, f3_ready = WireInit(false.B) 15709c6f1ddSLingrui98 1581d8f4dcbSJay fromFtq.req.ready := toICache(0).ready && toICache(1).ready && f2_ready && GTimer() > 500.U 15909c6f1ddSLingrui98 16037483030SJinYue toICache(0).valid := fromFtq.req.valid //&& !f0_flush 1611d8f4dcbSJay toICache(0).bits.vaddr := fromFtq.req.bits.startAddr 16237483030SJinYue toICache(1).valid := fromFtq.req.valid && f0_doubleLine //&& !f0_flush 16334a88126SJinYue toICache(1).bits.vaddr := fromFtq.req.bits.nextlineStart//fromFtq.req.bits.startAddr + (PredictWidth * 2).U //TODO: timing critical 16409c6f1ddSLingrui98 16558dbdfc2SJay /** <PERF> f0 fetch bubble */ 166f7c29b0aSJinYue 16758dbdfc2SJay XSPerfAccumulate("fetch_bubble_ftq_not_valid", !f0_valid ) 16858dbdfc2SJay XSPerfAccumulate("fetch_bubble_pipe_stall", f0_valid && toICache(0).ready && toICache(1).ready && !f1_ready ) 16958dbdfc2SJay XSPerfAccumulate("fetch_bubble_sram_0_busy", f0_valid && !toICache(0).ready ) 17058dbdfc2SJay XSPerfAccumulate("fetch_bubble_sram_1_busy", f0_valid && !toICache(1).ready ) 17158dbdfc2SJay 17258dbdfc2SJay 17358dbdfc2SJay /** 17458dbdfc2SJay ****************************************************************************** 17558dbdfc2SJay * IFU Stage 1 17658dbdfc2SJay * - calculate pc/half_pc/cut_ptr for every instruction 17758dbdfc2SJay ****************************************************************************** 17858dbdfc2SJay */ 17909c6f1ddSLingrui98 18009c6f1ddSLingrui98 val f1_valid = RegInit(false.B) 18109c6f1ddSLingrui98 val f1_ftq_req = RegEnable(next = f0_ftq_req, enable=f0_fire) 1826ce52296SJinYue // val f1_situation = RegEnable(next = f0_situation, enable=f0_fire) 18309c6f1ddSLingrui98 val f1_doubleLine = RegEnable(next = f0_doubleLine, enable=f0_fire) 18409c6f1ddSLingrui98 val f1_vSetIdx = RegEnable(next = f0_vSetIdx, enable=f0_fire) 1851d8f4dcbSJay val f1_fire = f1_valid && f1_ready 18609c6f1ddSLingrui98 1871d8f4dcbSJay f1_ready := f2_ready || !f1_valid 18809c6f1ddSLingrui98 1890d756c48SJinYue from_bpu_f1_flush := fromFtq.flushFromBpu.shouldFlushByStage3(f1_ftq_req.ftqIdx) && f1_valid 190cb4f77ceSLingrui98 // from_bpu_f1_flush := false.B 19109c6f1ddSLingrui98 19209c6f1ddSLingrui98 when(f1_flush) {f1_valid := false.B} 19309c6f1ddSLingrui98 .elsewhen(f0_fire && !f0_flush) {f1_valid := true.B} 19409c6f1ddSLingrui98 .elsewhen(f1_fire) {f1_valid := false.B} 19509c6f1ddSLingrui98 1962a3050c2SJay val f1_pc = VecInit((0 until PredictWidth).map(i => f1_ftq_req.startAddr + (i * 2).U)) 1972a3050c2SJay val f1_half_snpc = VecInit((0 until PredictWidth).map(i => f1_ftq_req.startAddr + ((i+2) * 2).U)) 1982a3050c2SJay val f1_cut_ptr = if(HasCExtension) VecInit((0 until PredictWidth + 1).map(i => Cat(0.U(1.W), f1_ftq_req.startAddr(blockOffBits-1, 1)) + i.U )) 1992a3050c2SJay else VecInit((0 until PredictWidth).map(i => Cat(0.U(1.W), f1_ftq_req.startAddr(blockOffBits-1, 2)) + i.U )) 20009c6f1ddSLingrui98 20158dbdfc2SJay /** 20258dbdfc2SJay ****************************************************************************** 20358dbdfc2SJay * IFU Stage 2 20458dbdfc2SJay * - icache response data (latched for pipeline stop) 20558dbdfc2SJay * - generate exceprion bits for every instruciton (page fault/access fault/mmio) 20658dbdfc2SJay * - generate predicted instruction range (1 means this instruciton is in this fetch packet) 20758dbdfc2SJay * - cut data from cachlines to packet instruction code 20858dbdfc2SJay * - instruction predecode and RVC expand 20958dbdfc2SJay ****************************************************************************** 21058dbdfc2SJay */ 21158dbdfc2SJay 2121d8f4dcbSJay val icacheRespAllValid = WireInit(false.B) 21309c6f1ddSLingrui98 21409c6f1ddSLingrui98 val f2_valid = RegInit(false.B) 21509c6f1ddSLingrui98 val f2_ftq_req = RegEnable(next = f1_ftq_req, enable=f1_fire) 2166ce52296SJinYue // val f2_situation = RegEnable(next = f1_situation, enable=f1_fire) 21709c6f1ddSLingrui98 val f2_doubleLine = RegEnable(next = f1_doubleLine, enable=f1_fire) 2181d8f4dcbSJay val f2_vSetIdx = RegEnable(next = f1_vSetIdx, enable=f1_fire) 2191d8f4dcbSJay val f2_fire = f2_valid && f2_ready 2201d8f4dcbSJay 2211d8f4dcbSJay f2_ready := f3_ready && icacheRespAllValid || !f2_valid 2221d8f4dcbSJay //TODO: addr compare may be timing critical 22334a88126SJinYue val f2_icache_all_resp_wire = fromICache(0).valid && (fromICache(0).bits.vaddr === f2_ftq_req.startAddr) && ((fromICache(1).valid && (fromICache(1).bits.vaddr === f2_ftq_req.nextlineStart)) || !f2_doubleLine) 2241d8f4dcbSJay val f2_icache_all_resp_reg = RegInit(false.B) 2251d8f4dcbSJay 2261d8f4dcbSJay icacheRespAllValid := f2_icache_all_resp_reg || f2_icache_all_resp_wire 2271d8f4dcbSJay 2281d8f4dcbSJay io.icacheStop := !f3_ready 2291d8f4dcbSJay 2301d8f4dcbSJay when(f2_flush) {f2_icache_all_resp_reg := false.B} 2311d8f4dcbSJay .elsewhen(f2_valid && f2_icache_all_resp_wire && !f3_ready) {f2_icache_all_resp_reg := true.B} 2321d8f4dcbSJay .elsewhen(f2_fire && f2_icache_all_resp_reg) {f2_icache_all_resp_reg := false.B} 23309c6f1ddSLingrui98 23409c6f1ddSLingrui98 when(f2_flush) {f2_valid := false.B} 23509c6f1ddSLingrui98 .elsewhen(f1_fire && !f1_flush) {f2_valid := true.B } 23609c6f1ddSLingrui98 .elsewhen(f2_fire) {f2_valid := false.B} 23709c6f1ddSLingrui98 2380bca1ccbSJinYue // val f2_cache_response_data = ResultHoldBypass(valid = f2_icache_all_resp_wire, data = VecInit(fromICache.map(_.bits.readData))) 2390bca1ccbSJinYue val f2_cache_response_data = VecInit(fromICache.map(_.bits.readData)) 2400bca1ccbSJinYue 24109c6f1ddSLingrui98 2421d8f4dcbSJay val f2_except_pf = VecInit((0 until PortNumber).map(i => fromICache(i).bits.tlbExcp.pageFault)) 2431d8f4dcbSJay val f2_except_af = VecInit((0 until PortNumber).map(i => fromICache(i).bits.tlbExcp.accessFault)) 244c0b2b8e9Srvcoresjw val f2_mmio = fromICache(0).bits.tlbExcp.mmio && !fromICache(0).bits.tlbExcp.accessFault && 245c0b2b8e9Srvcoresjw !fromICache(0).bits.tlbExcp.pageFault 2460be662e4SJay 2472a3050c2SJay val f2_pc = RegEnable(next = f1_pc, enable = f1_fire) 2482a3050c2SJay val f2_half_snpc = RegEnable(next = f1_half_snpc, enable = f1_fire) 2492a3050c2SJay val f2_cut_ptr = RegEnable(next = f1_cut_ptr, enable = f1_fire) 250a37fbf10SJay 251ee175d78SJay val f2_resend_vaddr = RegEnable(next = f1_ftq_req.startAddr + 2.U, enable = f1_fire) 2522a3050c2SJay 2532a3050c2SJay def isNextLine(pc: UInt, startAddr: UInt) = { 2542a3050c2SJay startAddr(blockOffBits) ^ pc(blockOffBits) 255b6982e83SLemover } 25609c6f1ddSLingrui98 2572a3050c2SJay def isLastInLine(pc: UInt) = { 2582a3050c2SJay pc(blockOffBits - 1, 0) === "b111110".U 25909c6f1ddSLingrui98 } 26009c6f1ddSLingrui98 2612a3050c2SJay val f2_foldpc = VecInit(f2_pc.map(i => XORFold(i(VAddrBits-1,1), MemPredPCWidth))) 2622a3050c2SJay val f2_jump_range = Fill(PredictWidth, !f2_ftq_req.ftqOffset.valid) | Fill(PredictWidth, 1.U(1.W)) >> ~f2_ftq_req.ftqOffset.bits 2631d011975SJinYue val f2_ftr_range = Fill(PredictWidth, f2_ftq_req.ftqOffset.valid) | Fill(PredictWidth, 1.U(1.W)) >> ~getBasicBlockIdx(f2_ftq_req.nextStartAddr, f2_ftq_req.startAddr) 2642a3050c2SJay val f2_instr_range = f2_jump_range & f2_ftr_range 2652a3050c2SJay val f2_pf_vec = VecInit((0 until PredictWidth).map(i => (!isNextLine(f2_pc(i), f2_ftq_req.startAddr) && f2_except_pf(0) || isNextLine(f2_pc(i), f2_ftq_req.startAddr) && f2_doubleLine && f2_except_pf(1)))) 2662a3050c2SJay val f2_af_vec = VecInit((0 until PredictWidth).map(i => (!isNextLine(f2_pc(i), f2_ftq_req.startAddr) && f2_except_af(0) || isNextLine(f2_pc(i), f2_ftq_req.startAddr) && f2_doubleLine && f2_except_af(1)))) 26709c6f1ddSLingrui98 2681d8f4dcbSJay val f2_paddrs = VecInit((0 until PortNumber).map(i => fromICache(i).bits.paddr)) 2691d8f4dcbSJay val f2_perf_info = io.icachePerfInfo 27009c6f1ddSLingrui98 2712a3050c2SJay def cut(cacheline: UInt, cutPtr: Vec[UInt]) : Vec[UInt] ={ 27209c6f1ddSLingrui98 if(HasCExtension){ 27309c6f1ddSLingrui98 val result = Wire(Vec(PredictWidth + 1, UInt(16.W))) 27409c6f1ddSLingrui98 val dataVec = cacheline.asTypeOf(Vec(blockBytes * 2/ 2, UInt(16.W))) 27509c6f1ddSLingrui98 (0 until PredictWidth + 1).foreach( i => 2762a3050c2SJay result(i) := dataVec(cutPtr(i)) 27709c6f1ddSLingrui98 ) 27809c6f1ddSLingrui98 result 27909c6f1ddSLingrui98 } else { 28009c6f1ddSLingrui98 val result = Wire(Vec(PredictWidth, UInt(32.W)) ) 28109c6f1ddSLingrui98 val dataVec = cacheline.asTypeOf(Vec(blockBytes * 2/ 4, UInt(32.W))) 28209c6f1ddSLingrui98 (0 until PredictWidth).foreach( i => 2832a3050c2SJay result(i) := dataVec(cutPtr(i)) 28409c6f1ddSLingrui98 ) 28509c6f1ddSLingrui98 result 28609c6f1ddSLingrui98 } 28709c6f1ddSLingrui98 } 28809c6f1ddSLingrui98 2892a3050c2SJay val f2_datas = VecInit((0 until PortNumber).map(i => f2_cache_response_data(i))) 2902a3050c2SJay val f2_cut_data = cut( Cat(f2_datas.map(cacheline => cacheline.asUInt ).reverse).asUInt, f2_cut_ptr ) 29109c6f1ddSLingrui98 29258dbdfc2SJay /** predecode (include RVC expander) */ 2932a3050c2SJay preDecoderIn.data := f2_cut_data 2942a3050c2SJay preDecoderIn.frontendTrigger := io.frontendTrigger 2952a3050c2SJay preDecoderIn.csrTriggerEnable := io.csrTriggerEnable 2962a3050c2SJay preDecoderIn.pc := f2_pc 29709c6f1ddSLingrui98 2982a3050c2SJay val f2_expd_instr = preDecoderOut.expInstr 2992a3050c2SJay val f2_pd = preDecoderOut.pd 3002a3050c2SJay val f2_jump_offset = preDecoderOut.jumpOffset 3012a3050c2SJay val f2_hasHalfValid = preDecoderOut.hasHalfValid 3022a3050c2SJay val f2_crossPageFault = VecInit((0 until PredictWidth).map(i => isLastInLine(f2_pc(i)) && !f2_except_pf(0) && f2_doubleLine && f2_except_pf(1) && !f2_pd(i).isRVC )) 30309c6f1ddSLingrui98 3041d8f4dcbSJay val predecodeOutValid = WireInit(false.B) 30509c6f1ddSLingrui98 30609c6f1ddSLingrui98 30758dbdfc2SJay /** 30858dbdfc2SJay ****************************************************************************** 30958dbdfc2SJay * IFU Stage 3 31058dbdfc2SJay * - handle MMIO instruciton 31158dbdfc2SJay * -send request to Uncache fetch Unit 31258dbdfc2SJay * -every packet include 1 MMIO instruction 31358dbdfc2SJay * -MMIO instructions will stop fetch pipeline until commiting from RoB 31458dbdfc2SJay * -flush to snpc (send ifu_redirect to Ftq) 31558dbdfc2SJay * - Ibuffer enqueue 31658dbdfc2SJay * - check predict result in Frontend (jalFault/retFault/notCFIFault/invalidTakenFault/targetFault) 31758dbdfc2SJay * - handle last half RVI instruction 31858dbdfc2SJay ****************************************************************************** 31958dbdfc2SJay */ 32058dbdfc2SJay 32109c6f1ddSLingrui98 val f3_valid = RegInit(false.B) 32209c6f1ddSLingrui98 val f3_ftq_req = RegEnable(next = f2_ftq_req, enable=f2_fire) 3236ce52296SJinYue // val f3_situation = RegEnable(next = f2_situation, enable=f2_fire) 32409c6f1ddSLingrui98 val f3_doubleLine = RegEnable(next = f2_doubleLine, enable=f2_fire) 3251d8f4dcbSJay val f3_fire = io.toIbuffer.fire() 3261d8f4dcbSJay 3271d8f4dcbSJay f3_ready := io.toIbuffer.ready || !f3_valid 32809c6f1ddSLingrui98 32909c6f1ddSLingrui98 val f3_cut_data = RegEnable(next = f2_cut_data, enable=f2_fire) 3301d8f4dcbSJay 33109c6f1ddSLingrui98 val f3_except_pf = RegEnable(next = f2_except_pf, enable = f2_fire) 33209c6f1ddSLingrui98 val f3_except_af = RegEnable(next = f2_except_af, enable = f2_fire) 3330be662e4SJay val f3_mmio = RegEnable(next = f2_mmio , enable = f2_fire) 33409c6f1ddSLingrui98 3352a3050c2SJay val f3_expd_instr = RegEnable(next = f2_expd_instr, enable = f2_fire) 3362a3050c2SJay val f3_pd = RegEnable(next = f2_pd, enable = f2_fire) 3372a3050c2SJay val f3_jump_offset = RegEnable(next = f2_jump_offset, enable = f2_fire) 3382a3050c2SJay val f3_af_vec = RegEnable(next = f2_af_vec, enable = f2_fire) 3392a3050c2SJay val f3_pf_vec = RegEnable(next = f2_pf_vec , enable = f2_fire) 3402a3050c2SJay val f3_pc = RegEnable(next = f2_pc, enable = f2_fire) 3412a3050c2SJay val f3_half_snpc = RegEnable(next = f2_half_snpc, enable = f2_fire) 3422a3050c2SJay val f3_instr_range = RegEnable(next = f2_instr_range, enable = f2_fire) 3432a3050c2SJay val f3_foldpc = RegEnable(next = f2_foldpc, enable = f2_fire) 3442a3050c2SJay val f3_crossPageFault = RegEnable(next = f2_crossPageFault, enable = f2_fire) 3452a3050c2SJay val f3_hasHalfValid = RegEnable(next = f2_hasHalfValid, enable = f2_fire) 34609c6f1ddSLingrui98 val f3_except = VecInit((0 until 2).map{i => f3_except_pf(i) || f3_except_af(i)}) 34709c6f1ddSLingrui98 val f3_has_except = f3_valid && (f3_except_af.reduce(_||_) || f3_except_pf.reduce(_||_)) 3481d8f4dcbSJay val f3_pAddrs = RegEnable(next = f2_paddrs, enable = f2_fire) 349ee175d78SJay val f3_resend_vaddr = RegEnable(next = f2_resend_vaddr, enable = f2_fire) 350ee175d78SJay 3511d011975SJinYue when(f3_valid && !f3_ftq_req.ftqOffset.valid){ 3521d011975SJinYue assert(f3_ftq_req.startAddr + 32.U >= f3_ftq_req.nextStartAddr , "More tha 32 Bytes fetch is not allowed!") 3531d011975SJinYue } 354a1351e5dSJay 3552a3050c2SJay /*** MMIO State Machine***/ 356ee175d78SJay val f3_mmio_data = Reg(Vec(2, UInt(16.W))) 357ee175d78SJay val mmio_is_RVC = RegInit(false.B) 358ee175d78SJay val mmio_resend_addr =RegInit(0.U(PAddrBits.W)) 359ee175d78SJay val mmio_resend_af = RegInit(false.B) 360a37fbf10SJay 361ee175d78SJay val m_idle :: m_sendReq :: m_waitResp :: m_sendTLB :: m_tlbResp :: m_sendPMP :: m_resendReq :: m_waitResendResp :: m_waitCommit :: m_commited :: Nil = Enum(10) 362ee175d78SJay val mmio_state = RegInit(m_idle) 363a37fbf10SJay 3649bae7d6eSJay val f3_req_is_mmio = f3_mmio && f3_valid 3652a3050c2SJay val mmio_commit = VecInit(io.rob_commits.map{commit => commit.valid && commit.bits.ftqIdx === f3_ftq_req.ftqIdx && commit.bits.ftqOffset === 0.U}).asUInt.orR 366ee175d78SJay val f3_mmio_req_commit = f3_req_is_mmio && mmio_state === m_commited 367a37fbf10SJay 368ee175d78SJay val f3_mmio_to_commit = f3_req_is_mmio && mmio_state === m_waitCommit 369a37fbf10SJay val f3_mmio_to_commit_next = RegNext(f3_mmio_to_commit) 370a37fbf10SJay val f3_mmio_can_go = f3_mmio_to_commit && !f3_mmio_to_commit_next 371a37fbf10SJay 37256788a33SJinYue val fromFtqRedirectReg = RegNext(fromFtq.redirect) 37356788a33SJinYue val f3_ftq_flush_self = fromFtqRedirectReg.valid && RedirectLevel.flushItself(fromFtqRedirectReg.bits.level) 37456788a33SJinYue val f3_ftq_flush_by_older = fromFtqRedirectReg.valid && isBefore(fromFtqRedirectReg.bits.ftqIdx, f3_ftq_req.ftqIdx) 3759bae7d6eSJay 37656788a33SJinYue val f3_need_not_flush = f3_req_is_mmio && fromFtqRedirectReg.valid && !f3_ftq_flush_self && !f3_ftq_flush_by_older 3779bae7d6eSJay 3789bae7d6eSJay when(f3_flush && !f3_need_not_flush) {f3_valid := false.B} 379a37fbf10SJay .elsewhen(f2_fire && !f2_flush ) {f3_valid := true.B } 380a37fbf10SJay .elsewhen(io.toIbuffer.fire() && !f3_req_is_mmio) {f3_valid := false.B} 381a37fbf10SJay .elsewhen{f3_req_is_mmio && f3_mmio_req_commit} {f3_valid := false.B} 382a37fbf10SJay 383a37fbf10SJay val f3_mmio_use_seq_pc = RegInit(false.B) 384a37fbf10SJay 38556788a33SJinYue val (redirect_ftqIdx, redirect_ftqOffset) = (fromFtqRedirectReg.bits.ftqIdx,fromFtqRedirectReg.bits.ftqOffset) 38656788a33SJinYue val redirect_mmio_req = fromFtqRedirectReg.valid && redirect_ftqIdx === f3_ftq_req.ftqIdx && redirect_ftqOffset === 0.U 387a37fbf10SJay 388a37fbf10SJay when(RegNext(f2_fire && !f2_flush) && f3_req_is_mmio) { f3_mmio_use_seq_pc := true.B } 389a37fbf10SJay .elsewhen(redirect_mmio_req) { f3_mmio_use_seq_pc := false.B } 390a37fbf10SJay 391a37fbf10SJay f3_ready := Mux(f3_req_is_mmio, io.toIbuffer.ready && f3_mmio_req_commit || !f3_valid , io.toIbuffer.ready || !f3_valid) 392a37fbf10SJay 393ee175d78SJay // when(fromUncache.fire()) {f3_mmio_data := fromUncache.bits.data} 394a37fbf10SJay 395a37fbf10SJay 396a37fbf10SJay switch(mmio_state){ 397ee175d78SJay is(m_idle){ 3989bae7d6eSJay when(f3_req_is_mmio){ 399ee175d78SJay mmio_state := m_sendReq 400a37fbf10SJay } 401a37fbf10SJay } 402a37fbf10SJay 403ee175d78SJay is(m_sendReq){ 404ee175d78SJay mmio_state := Mux(toUncache.fire(), m_waitResp, m_sendReq ) 405a37fbf10SJay } 406a37fbf10SJay 407ee175d78SJay is(m_waitResp){ 408a37fbf10SJay when(fromUncache.fire()){ 409a37fbf10SJay val isRVC = fromUncache.bits.data(1,0) =/= 3.U 410ee175d78SJay val needResend = !isRVC && f3_pAddrs(0)(2,1) === 3.U 411ee175d78SJay mmio_state := Mux(needResend, m_sendTLB , m_waitCommit) 412ee175d78SJay 413ee175d78SJay mmio_is_RVC := isRVC 414ee175d78SJay f3_mmio_data(0) := fromUncache.bits.data(15,0) 415ee175d78SJay f3_mmio_data(1) := fromUncache.bits.data(31,16) 416a37fbf10SJay } 417a37fbf10SJay } 418a37fbf10SJay 419ee175d78SJay is(m_sendTLB){ 420ee175d78SJay mmio_state := m_tlbResp 421a37fbf10SJay } 422a37fbf10SJay 423ee175d78SJay is(m_tlbResp){ 424ee175d78SJay mmio_state := m_sendPMP 425ee175d78SJay mmio_resend_addr := io.iTLBInter.resp.bits.paddr 426ee175d78SJay } 427ee175d78SJay 428ee175d78SJay is(m_sendPMP){ 429ee175d78SJay val pmpExcpAF = io.pmp.resp.instr 430ee175d78SJay mmio_state := Mux(pmpExcpAF, m_waitCommit , m_resendReq) 431ee175d78SJay mmio_resend_af := pmpExcpAF 432ee175d78SJay } 433ee175d78SJay 434ee175d78SJay is(m_resendReq){ 435ee175d78SJay mmio_state := Mux(toUncache.fire(), m_waitResendResp, m_resendReq ) 436ee175d78SJay } 437ee175d78SJay 438ee175d78SJay is(m_waitResendResp){ 439a37fbf10SJay when(fromUncache.fire()){ 440ee175d78SJay mmio_state := m_waitCommit 441ee175d78SJay f3_mmio_data(1) := fromUncache.bits.data(15,0) 442a37fbf10SJay } 443a37fbf10SJay } 444a37fbf10SJay 445ee175d78SJay is(m_waitCommit){ 4462a3050c2SJay when(mmio_commit){ 447ee175d78SJay mmio_state := m_commited 448a37fbf10SJay } 449a37fbf10SJay } 4502a3050c2SJay 451ee175d78SJay //normal mmio instruction 452ee175d78SJay is(m_commited){ 453ee175d78SJay mmio_state := m_idle 454ee175d78SJay mmio_is_RVC := false.B 455ee175d78SJay mmio_resend_addr := 0.U 4562a3050c2SJay } 457a37fbf10SJay } 458a37fbf10SJay 459ee175d78SJay //exception or flush by older branch prediction 460167bcd01SJay when(f3_ftq_flush_self || f3_ftq_flush_by_older) { 461ee175d78SJay mmio_state := m_idle 462ee175d78SJay mmio_is_RVC := false.B 463ee175d78SJay mmio_resend_addr := 0.U 464ee175d78SJay mmio_resend_af := false.B 465ee175d78SJay f3_mmio_data.map(_ := 0.U) 4669bae7d6eSJay } 4679bae7d6eSJay 468ee175d78SJay toUncache.valid := ((mmio_state === m_sendReq) || (mmio_state === m_resendReq)) && f3_req_is_mmio 469ee175d78SJay toUncache.bits.addr := Mux((mmio_state === m_resendReq), mmio_resend_addr, f3_pAddrs(0)) 470a37fbf10SJay fromUncache.ready := true.B 471a37fbf10SJay 472ee175d78SJay io.iTLBInter.req.valid := (mmio_state === m_sendTLB) && f3_req_is_mmio 473ee175d78SJay io.iTLBInter.req.bits.size := 3.U 474ee175d78SJay io.iTLBInter.req.bits.vaddr := f3_resend_vaddr 475ee175d78SJay io.iTLBInter.req.bits.debug.pc := f3_resend_vaddr 476ee175d78SJay 477ee175d78SJay io.iTLBInter.req.bits.cmd := TlbCmd.exec 478ee175d78SJay io.iTLBInter.req.bits.robIdx := DontCare 479ee175d78SJay io.iTLBInter.req.bits.debug.isFirstIssue := DontCare 480ee175d78SJay 481ee175d78SJay io.pmp.req.valid := (mmio_state === m_sendPMP) && f3_req_is_mmio 482ee175d78SJay io.pmp.req.bits.addr := mmio_resend_addr 483ee175d78SJay io.pmp.req.bits.size := 3.U 484ee175d78SJay io.pmp.req.bits.cmd := TlbCmd.exec 485f7c29b0aSJinYue 4862a3050c2SJay val f3_lastHalf = RegInit(0.U.asTypeOf(new LastHalfInfo)) 48709c6f1ddSLingrui98 48809c6f1ddSLingrui98 val f3_predecode_range = VecInit(preDecoderOut.pd.map(inst => inst.valid)).asUInt 4890be662e4SJay val f3_mmio_range = VecInit((0 until PredictWidth).map(i => if(i ==0) true.B else false.B)) 4902a3050c2SJay val f3_instr_valid = Wire(Vec(PredictWidth, Bool())) 49109c6f1ddSLingrui98 4922a3050c2SJay /*** prediction result check ***/ 4932a3050c2SJay checkerIn.ftqOffset := f3_ftq_req.ftqOffset 4942a3050c2SJay checkerIn.jumpOffset := f3_jump_offset 4956ce52296SJinYue checkerIn.target := f3_ftq_req.nextStartAddr 4962a3050c2SJay checkerIn.instrRange := f3_instr_range.asTypeOf(Vec(PredictWidth, Bool())) 4972a3050c2SJay checkerIn.instrValid := f3_instr_valid.asTypeOf(Vec(PredictWidth, Bool())) 4982a3050c2SJay checkerIn.pds := f3_pd 4992a3050c2SJay checkerIn.pc := f3_pc 5002a3050c2SJay 50158dbdfc2SJay /*** handle half RVI in the last 2 Bytes ***/ 5022a3050c2SJay 5032a3050c2SJay def hasLastHalf(idx: UInt) = { 5041d011975SJinYue !f3_pd(idx).isRVC && checkerOut.fixedRange(idx) && f3_instr_valid(idx) && !checkerOut.fixedTaken(idx) && !checkerOut.fixedMissPred(idx) && ! f3_req_is_mmio 5052a3050c2SJay } 5062a3050c2SJay 5072a3050c2SJay val f3_last_validIdx = ~ParallelPriorityEncoder(checkerOut.fixedRange.reverse) 5082a3050c2SJay 5092a3050c2SJay val f3_hasLastHalf = hasLastHalf((PredictWidth - 1).U) 5102a3050c2SJay val f3_false_lastHalf = hasLastHalf(f3_last_validIdx) 5112a3050c2SJay val f3_false_snpc = f3_half_snpc(f3_last_validIdx) 5122a3050c2SJay 5132a3050c2SJay val f3_lastHalf_mask = VecInit((0 until PredictWidth).map( i => if(i ==0) false.B else true.B )).asUInt() 5142a3050c2SJay 5152a3050c2SJay when (f3_flush) { 5162a3050c2SJay f3_lastHalf.valid := false.B 5172a3050c2SJay }.elsewhen (f3_fire) { 5182a3050c2SJay f3_lastHalf.valid := f3_hasLastHalf 5196ce52296SJinYue f3_lastHalf.middlePC := f3_ftq_req.nextStartAddr 5202a3050c2SJay } 5212a3050c2SJay 5222a3050c2SJay f3_instr_valid := Mux(f3_lastHalf.valid,f3_hasHalfValid ,VecInit(f3_pd.map(inst => inst.valid))) 5232a3050c2SJay 5242a3050c2SJay /*** frontend Trigger ***/ 5252a3050c2SJay frontendTrigger.io.pds := f3_pd 5262a3050c2SJay frontendTrigger.io.pc := f3_pc 5272a3050c2SJay frontendTrigger.io.data := f3_cut_data 5282a3050c2SJay 5292a3050c2SJay frontendTrigger.io.frontendTrigger := io.frontendTrigger 5302a3050c2SJay frontendTrigger.io.csrTriggerEnable := io.csrTriggerEnable 5312a3050c2SJay 5322a3050c2SJay val f3_triggered = frontendTrigger.io.triggered 5332a3050c2SJay 5342a3050c2SJay /*** send to Ibuffer ***/ 5352a3050c2SJay 5362a3050c2SJay io.toIbuffer.valid := f3_valid && (!f3_req_is_mmio || f3_mmio_can_go) && !f3_flush 5372a3050c2SJay io.toIbuffer.bits.instrs := f3_expd_instr 5382a3050c2SJay io.toIbuffer.bits.valid := f3_instr_valid.asUInt 5392a3050c2SJay io.toIbuffer.bits.enqEnable := checkerOut.fixedRange.asUInt & f3_instr_valid.asUInt 5402a3050c2SJay io.toIbuffer.bits.pd := f3_pd 54109c6f1ddSLingrui98 io.toIbuffer.bits.ftqPtr := f3_ftq_req.ftqIdx 5422a3050c2SJay io.toIbuffer.bits.pc := f3_pc 5432a3050c2SJay io.toIbuffer.bits.ftqOffset.zipWithIndex.map{case(a, i) => a.bits := i.U; a.valid := checkerOut.fixedTaken(i) && !f3_req_is_mmio} 5442a3050c2SJay io.toIbuffer.bits.foldpc := f3_foldpc 5453908fff2SJay io.toIbuffer.bits.ipf := VecInit(f3_pf_vec.zip(f3_crossPageFault).map{case (pf, crossPF) => pf || crossPF}) 5462a3050c2SJay io.toIbuffer.bits.acf := f3_af_vec 5472a3050c2SJay io.toIbuffer.bits.crossPageIPFFix := f3_crossPageFault 5482a3050c2SJay io.toIbuffer.bits.triggered := f3_triggered 5492a3050c2SJay 5502a3050c2SJay val lastHalfMask = VecInit((0 until PredictWidth).map(i => if(i ==0) false.B else true.B)) 5512a3050c2SJay when(f3_lastHalf.valid){ 5522a3050c2SJay io.toIbuffer.bits.enqEnable := checkerOut.fixedRange.asUInt & f3_instr_valid.asUInt & lastHalfMask.asUInt 5532a3050c2SJay io.toIbuffer.bits.valid := f3_lastHalf_mask & f3_instr_valid.asUInt 5542a3050c2SJay } 5552a3050c2SJay 5562a3050c2SJay /** external predecode for MMIO instruction */ 5572a3050c2SJay when(f3_req_is_mmio){ 558ee175d78SJay val inst = Cat(f3_mmio_data(1), f3_mmio_data(0)) 5592a3050c2SJay val currentIsRVC = isRVC(inst) 5602a3050c2SJay 5612a3050c2SJay val brType::isCall::isRet::Nil = brInfo(inst) 5622a3050c2SJay val jalOffset = jal_offset(inst, currentIsRVC) 5632a3050c2SJay val brOffset = br_offset(inst, currentIsRVC) 5642a3050c2SJay 5652a3050c2SJay io.toIbuffer.bits.instrs (0) := new RVCDecoder(inst, XLEN).decode.bits 5662a3050c2SJay 5672a3050c2SJay io.toIbuffer.bits.pd(0).valid := true.B 5682a3050c2SJay io.toIbuffer.bits.pd(0).isRVC := currentIsRVC 5692a3050c2SJay io.toIbuffer.bits.pd(0).brType := brType 5702a3050c2SJay io.toIbuffer.bits.pd(0).isCall := isCall 5712a3050c2SJay io.toIbuffer.bits.pd(0).isRet := isRet 5722a3050c2SJay 573ee175d78SJay io.toIbuffer.bits.acf(0) := mmio_resend_af 574ee175d78SJay 5752a3050c2SJay io.toIbuffer.bits.enqEnable := f3_mmio_range.asUInt 5762a3050c2SJay } 5772a3050c2SJay 57809c6f1ddSLingrui98 57909c6f1ddSLingrui98 //Write back to Ftq 580a37fbf10SJay val f3_cache_fetch = f3_valid && !(f2_fire && !f2_flush) 581a37fbf10SJay val finishFetchMaskReg = RegNext(f3_cache_fetch) 582a37fbf10SJay 5832a3050c2SJay val mmioFlushWb = Wire(Valid(new PredecodeWritebackBundle)) 5840be662e4SJay val f3_mmio_missOffset = Wire(ValidUndirectioned(UInt(log2Ceil(PredictWidth).W))) 585a37fbf10SJay f3_mmio_missOffset.valid := f3_req_is_mmio 5860be662e4SJay f3_mmio_missOffset.bits := 0.U 5870be662e4SJay 588ee175d78SJay mmioFlushWb.valid := (f3_req_is_mmio && mmio_state === m_waitCommit && RegNext(fromUncache.fire()) && f3_mmio_use_seq_pc) 5892a3050c2SJay mmioFlushWb.bits.pc := f3_pc 5902a3050c2SJay mmioFlushWb.bits.pd := f3_pd 5912a3050c2SJay mmioFlushWb.bits.pd.zipWithIndex.map{case(instr,i) => instr.valid := f3_mmio_range(i)} 5922a3050c2SJay mmioFlushWb.bits.ftqIdx := f3_ftq_req.ftqIdx 5932a3050c2SJay mmioFlushWb.bits.ftqOffset := f3_ftq_req.ftqOffset.bits 5942a3050c2SJay mmioFlushWb.bits.misOffset := f3_mmio_missOffset 5952a3050c2SJay mmioFlushWb.bits.cfiOffset := DontCare 596ee175d78SJay mmioFlushWb.bits.target := Mux(mmio_is_RVC, f3_ftq_req.startAddr + 2.U , f3_ftq_req.startAddr + 4.U) 5972a3050c2SJay mmioFlushWb.bits.jalTarget := DontCare 5982a3050c2SJay mmioFlushWb.bits.instrRange := f3_mmio_range 59909c6f1ddSLingrui98 600ee175d78SJay mmio_redirect := (f3_req_is_mmio && mmio_state === m_waitCommit && RegNext(fromUncache.fire()) && f3_mmio_use_seq_pc) 60109c6f1ddSLingrui98 60258dbdfc2SJay /** 60358dbdfc2SJay ****************************************************************************** 60458dbdfc2SJay * IFU Write Back Stage 60558dbdfc2SJay * - write back predecode information to Ftq to update 60658dbdfc2SJay * - redirect if found fault prediction 60758dbdfc2SJay * - redirect if has false hit last half (last PC is not start + 32 Bytes, but in the midle of an notCFI RVI instruction) 60858dbdfc2SJay ****************************************************************************** 6092a3050c2SJay */ 61058dbdfc2SJay 6112a3050c2SJay val wb_valid = RegNext(RegNext(f2_fire && !f2_flush) && !f3_req_is_mmio && !f3_flush) 6122a3050c2SJay val wb_ftq_req = RegNext(f3_ftq_req) 613cd365d4cSrvcoresjw 6142a3050c2SJay val wb_check_result = RegNext(checkerOut) 6152a3050c2SJay val wb_instr_range = RegNext(io.toIbuffer.bits.enqEnable) 6162a3050c2SJay val wb_pc = RegNext(f3_pc) 6172a3050c2SJay val wb_pd = RegNext(f3_pd) 6182a3050c2SJay val wb_instr_valid = RegNext(f3_instr_valid) 6192a3050c2SJay 6202a3050c2SJay /* false hit lastHalf */ 6212a3050c2SJay val wb_lastIdx = RegNext(f3_last_validIdx) 6222a3050c2SJay val wb_false_lastHalf = RegNext(f3_false_lastHalf) && wb_lastIdx =/= (PredictWidth - 1).U 6232a3050c2SJay val wb_false_target = RegNext(f3_false_snpc) 6242a3050c2SJay 6252a3050c2SJay val wb_half_flush = wb_false_lastHalf 6262a3050c2SJay val wb_half_target = wb_false_target 6272a3050c2SJay 628a1351e5dSJay /* false oversize */ 629a1351e5dSJay val lastIsRVC = wb_instr_range.asTypeOf(Vec(PredictWidth,Bool())).last && wb_pd.last.isRVC 630a1351e5dSJay val lastIsRVI = wb_instr_range.asTypeOf(Vec(PredictWidth,Bool()))(PredictWidth - 2) && !wb_pd(PredictWidth - 2).isRVC 631a1351e5dSJay val lastTaken = wb_check_result.fixedTaken.last 632a1351e5dSJay 6332a3050c2SJay f3_wb_not_flush := wb_ftq_req.ftqIdx === f3_ftq_req.ftqIdx && f3_valid && wb_valid 6342a3050c2SJay 6352a3050c2SJay val checkFlushWb = Wire(Valid(new PredecodeWritebackBundle)) 6362a3050c2SJay checkFlushWb.valid := wb_valid 6372a3050c2SJay checkFlushWb.bits.pc := wb_pc 6382a3050c2SJay checkFlushWb.bits.pd := wb_pd 6392a3050c2SJay checkFlushWb.bits.pd.zipWithIndex.map{case(instr,i) => instr.valid := wb_instr_valid(i)} 6402a3050c2SJay checkFlushWb.bits.ftqIdx := wb_ftq_req.ftqIdx 6412a3050c2SJay checkFlushWb.bits.ftqOffset := wb_ftq_req.ftqOffset.bits 6421d011975SJinYue checkFlushWb.bits.misOffset.valid := ParallelOR(wb_check_result.fixedMissPred) || wb_half_flush 6432a3050c2SJay checkFlushWb.bits.misOffset.bits := Mux(wb_half_flush, (PredictWidth - 1).U, ParallelPriorityEncoder(wb_check_result.fixedMissPred)) 6442a3050c2SJay checkFlushWb.bits.cfiOffset.valid := ParallelOR(wb_check_result.fixedTaken) 6452a3050c2SJay checkFlushWb.bits.cfiOffset.bits := ParallelPriorityEncoder(wb_check_result.fixedTaken) 6461d011975SJinYue checkFlushWb.bits.target := Mux(wb_half_flush, wb_half_target, wb_check_result.fixedTarget(ParallelPriorityEncoder(wb_check_result.fixedMissPred))) 647b37e4b45SLingrui98 checkFlushWb.bits.jalTarget := wb_check_result.fixedTarget(ParallelPriorityEncoder(VecInit(wb_pd.zip(wb_instr_valid).map{case (pd, v) => v && pd.isJal }))) 6482a3050c2SJay checkFlushWb.bits.instrRange := wb_instr_range.asTypeOf(Vec(PredictWidth, Bool())) 6492a3050c2SJay 6502a3050c2SJay toFtq.pdWb := Mux(f3_req_is_mmio, mmioFlushWb, checkFlushWb) 6512a3050c2SJay 6522a3050c2SJay wb_redirect := checkFlushWb.bits.misOffset.valid && wb_valid 65309c6f1ddSLingrui98 6541d8f4dcbSJay 655*5b3c20f7SJinYue /*write back flush type*/ 656*5b3c20f7SJinYue val checkFaultType = wb_check_result.faultType 657*5b3c20f7SJinYue val checkJalFault = wb_valid && checkFaultType.map(_.isjalFault).reduce(_||_) 658*5b3c20f7SJinYue val checkRetFault = wb_valid && checkFaultType.map(_.isRetFault).reduce(_||_) 659*5b3c20f7SJinYue val checkTargetFault = wb_valid && checkFaultType.map(_.istargetFault).reduce(_||_) 660*5b3c20f7SJinYue val checkNotCFIFault = wb_valid && checkFaultType.map(_.notCFIFault).reduce(_||_) 661*5b3c20f7SJinYue val checkInvalidTaken = wb_valid && checkFaultType.map(_.invalidTakenFault).reduce(_||_) 662*5b3c20f7SJinYue 663*5b3c20f7SJinYue 664*5b3c20f7SJinYue XSPerfAccumulate("predecode_flush_jalFault", checkJalFault ) 665*5b3c20f7SJinYue XSPerfAccumulate("predecode_flush_retFault", checkRetFault ) 666*5b3c20f7SJinYue XSPerfAccumulate("predecode_flush_targetFault", checkTargetFault ) 667*5b3c20f7SJinYue XSPerfAccumulate("predecode_flush_notCFIFault", checkNotCFIFault ) 668*5b3c20f7SJinYue XSPerfAccumulate("predecode_flush_incalidTakenFault", checkInvalidTaken ) 669*5b3c20f7SJinYue 670*5b3c20f7SJinYue when(checkRetFault){ 671*5b3c20f7SJinYue XSDebug("startAddr:%x nextstartAddr:%x taken:%d takenIdx:%d\n", 672*5b3c20f7SJinYue wb_ftq_req.startAddr, wb_ftq_req.nextStartAddr, wb_ftq_req.ftqOffset.valid, wb_ftq_req.ftqOffset.bits) 673*5b3c20f7SJinYue } 674*5b3c20f7SJinYue 6751d8f4dcbSJay /** performance counter */ 6761d8f4dcbSJay val f3_perf_info = RegEnable(next = f2_perf_info, enable = f2_fire) 6771d8f4dcbSJay val f3_req_0 = io.toIbuffer.fire() 6781d8f4dcbSJay val f3_req_1 = io.toIbuffer.fire() && f3_doubleLine 6791d8f4dcbSJay val f3_hit_0 = io.toIbuffer.fire() && f3_perf_info.bank_hit(0) 6801d8f4dcbSJay val f3_hit_1 = io.toIbuffer.fire() && f3_doubleLine & f3_perf_info.bank_hit(1) 6811d8f4dcbSJay val f3_hit = f3_perf_info.hit 682cd365d4cSrvcoresjw val perfEvents = Seq( 6832a3050c2SJay ("frontendFlush ", wb_redirect ), 684cd365d4cSrvcoresjw ("ifu_req ", io.toIbuffer.fire() ), 6851d8f4dcbSJay ("ifu_miss ", io.toIbuffer.fire() && !f3_perf_info.hit ), 686cd365d4cSrvcoresjw ("ifu_req_cacheline_0 ", f3_req_0 ), 687cd365d4cSrvcoresjw ("ifu_req_cacheline_1 ", f3_req_1 ), 688cd365d4cSrvcoresjw ("ifu_req_cacheline_0_hit ", f3_hit_1 ), 689cd365d4cSrvcoresjw ("ifu_req_cacheline_1_hit ", f3_hit_1 ), 6901d8f4dcbSJay ("only_0_hit ", f3_perf_info.only_0_hit && io.toIbuffer.fire() ), 6911d8f4dcbSJay ("only_0_miss ", f3_perf_info.only_0_miss && io.toIbuffer.fire() ), 6921d8f4dcbSJay ("hit_0_hit_1 ", f3_perf_info.hit_0_hit_1 && io.toIbuffer.fire() ), 6931d8f4dcbSJay ("hit_0_miss_1 ", f3_perf_info.hit_0_miss_1 && io.toIbuffer.fire() ), 6941d8f4dcbSJay ("miss_0_hit_1 ", f3_perf_info.miss_0_hit_1 && io.toIbuffer.fire() ), 6951d8f4dcbSJay ("miss_0_miss_1 ", f3_perf_info.miss_0_miss_1 && io.toIbuffer.fire() ), 6966ce52296SJinYue // ("cross_line_block ", io.toIbuffer.fire() && f3_situation(0) ), 6976ce52296SJinYue // ("fall_through_is_cacheline_end", io.toIbuffer.fire() && f3_situation(1) ), 698cd365d4cSrvcoresjw ) 6991ca0e4f3SYinan Xu generatePerfEvent() 70009c6f1ddSLingrui98 701f7c29b0aSJinYue XSPerfAccumulate("ifu_req", io.toIbuffer.fire() ) 702f7c29b0aSJinYue XSPerfAccumulate("ifu_miss", io.toIbuffer.fire() && !f3_hit ) 703f7c29b0aSJinYue XSPerfAccumulate("ifu_req_cacheline_0", f3_req_0 ) 704f7c29b0aSJinYue XSPerfAccumulate("ifu_req_cacheline_1", f3_req_1 ) 705f7c29b0aSJinYue XSPerfAccumulate("ifu_req_cacheline_0_hit", f3_hit_0 ) 706f7c29b0aSJinYue XSPerfAccumulate("ifu_req_cacheline_1_hit", f3_hit_1 ) 7072a3050c2SJay XSPerfAccumulate("frontendFlush", wb_redirect ) 7081d8f4dcbSJay XSPerfAccumulate("only_0_hit", f3_perf_info.only_0_hit && io.toIbuffer.fire() ) 7091d8f4dcbSJay XSPerfAccumulate("only_0_miss", f3_perf_info.only_0_miss && io.toIbuffer.fire() ) 7101d8f4dcbSJay XSPerfAccumulate("hit_0_hit_1", f3_perf_info.hit_0_hit_1 && io.toIbuffer.fire() ) 7111d8f4dcbSJay XSPerfAccumulate("hit_0_miss_1", f3_perf_info.hit_0_miss_1 && io.toIbuffer.fire() ) 7121d8f4dcbSJay XSPerfAccumulate("miss_0_hit_1", f3_perf_info.miss_0_hit_1 && io.toIbuffer.fire() ) 7131d8f4dcbSJay XSPerfAccumulate("miss_0_miss_1", f3_perf_info.miss_0_miss_1 && io.toIbuffer.fire() ) 714a108d429SJay XSPerfAccumulate("hit_0_except_1", f3_perf_info.hit_0_except_1 && io.toIbuffer.fire() ) 715a108d429SJay XSPerfAccumulate("miss_0_except_1", f3_perf_info.miss_0_except_1 && io.toIbuffer.fire() ) 716a108d429SJay XSPerfAccumulate("except_0", f3_perf_info.except_0 && io.toIbuffer.fire() ) 71709c6f1ddSLingrui98} 718