109c6f1ddSLingrui98/*************************************************************************************** 209c6f1ddSLingrui98* Copyright (c) 2020-2021 Institute of Computing Technology, Chinese Academy of Sciences 309c6f1ddSLingrui98* Copyright (c) 2020-2021 Peng Cheng Laboratory 409c6f1ddSLingrui98* 509c6f1ddSLingrui98* XiangShan is licensed under Mulan PSL v2. 609c6f1ddSLingrui98* You can use this software according to the terms and conditions of the Mulan PSL v2. 709c6f1ddSLingrui98* You may obtain a copy of Mulan PSL v2 at: 809c6f1ddSLingrui98* http://license.coscl.org.cn/MulanPSL2 909c6f1ddSLingrui98* 1009c6f1ddSLingrui98* THIS SOFTWARE IS PROVIDED ON AN "AS IS" BASIS, WITHOUT WARRANTIES OF ANY KIND, 1109c6f1ddSLingrui98* EITHER EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO NON-INFRINGEMENT, 1209c6f1ddSLingrui98* MERCHANTABILITY OR FIT FOR A PARTICULAR PURPOSE. 1309c6f1ddSLingrui98* 1409c6f1ddSLingrui98* See the Mulan PSL v2 for more details. 1509c6f1ddSLingrui98***************************************************************************************/ 1609c6f1ddSLingrui98 1709c6f1ddSLingrui98package xiangshan.frontend 1809c6f1ddSLingrui98 1909c6f1ddSLingrui98import chipsalliance.rocketchip.config.Parameters 2009c6f1ddSLingrui98import chisel3._ 2109c6f1ddSLingrui98import chisel3.util._ 222a3050c2SJayimport freechips.rocketchip.rocket.RVCDecoder 2309c6f1ddSLingrui98import xiangshan._ 2409c6f1ddSLingrui98import xiangshan.cache.mmu._ 251d8f4dcbSJayimport xiangshan.frontend.icache._ 2609c6f1ddSLingrui98import utils._ 27b6982e83SLemoverimport xiangshan.backend.fu.{PMPReqBundle, PMPRespBundle} 2809c6f1ddSLingrui98 2909c6f1ddSLingrui98trait HasInstrMMIOConst extends HasXSParameter with HasIFUConst{ 3009c6f1ddSLingrui98 def mmioBusWidth = 64 3109c6f1ddSLingrui98 def mmioBusBytes = mmioBusWidth / 8 320be662e4SJay def maxInstrLen = 32 3309c6f1ddSLingrui98} 3409c6f1ddSLingrui98 3509c6f1ddSLingrui98trait HasIFUConst extends HasXSParameter{ 361d8f4dcbSJay def addrAlign(addr: UInt, bytes: Int, highest: Int): UInt = Cat(addr(highest-1, log2Ceil(bytes)), 0.U(log2Ceil(bytes).W)) 371d8f4dcbSJay def fetchQueueSize = 2 381d8f4dcbSJay 392a3050c2SJay def getBasicBlockIdx( pc: UInt, start: UInt ): UInt = { 402a3050c2SJay val byteOffset = pc - start 412a3050c2SJay (byteOffset - instBytes.U)(log2Ceil(PredictWidth),instOffsetBits) 421d8f4dcbSJay } 4309c6f1ddSLingrui98} 4409c6f1ddSLingrui98 4509c6f1ddSLingrui98class IfuToFtqIO(implicit p:Parameters) extends XSBundle { 4609c6f1ddSLingrui98 val pdWb = Valid(new PredecodeWritebackBundle) 4709c6f1ddSLingrui98} 4809c6f1ddSLingrui98 4909c6f1ddSLingrui98class FtqInterface(implicit p: Parameters) extends XSBundle { 5009c6f1ddSLingrui98 val fromFtq = Flipped(new FtqToIfuIO) 5109c6f1ddSLingrui98 val toFtq = new IfuToFtqIO 5209c6f1ddSLingrui98} 5309c6f1ddSLingrui98 540be662e4SJayclass UncacheInterface(implicit p: Parameters) extends XSBundle { 550be662e4SJay val fromUncache = Flipped(DecoupledIO(new InsUncacheResp)) 560be662e4SJay val toUncache = DecoupledIO( new InsUncacheReq ) 570be662e4SJay} 5809c6f1ddSLingrui98class NewIFUIO(implicit p: Parameters) extends XSBundle { 5909c6f1ddSLingrui98 val ftqInter = new FtqInterface 6050780602SJenius val icacheInter = Flipped(new IFUICacheIO) 611d8f4dcbSJay val icacheStop = Output(Bool()) 621d8f4dcbSJay val icachePerfInfo = Input(new ICachePerfInfo) 6309c6f1ddSLingrui98 val toIbuffer = Decoupled(new FetchToIBuffer) 640be662e4SJay val uncacheInter = new UncacheInterface 6572951335SLi Qianruo val frontendTrigger = Flipped(new FrontendTdataDistributeIO) 6672951335SLi Qianruo val csrTriggerEnable = Input(Vec(4, Bool())) 67a37fbf10SJay val rob_commits = Flipped(Vec(CommitWidth, Valid(new RobCommitInfo))) 68f1fe8698SLemover val iTLBInter = new TlbRequestIO 6956788a33SJinYue val pmp = new ICachePMPBundle 7009c6f1ddSLingrui98} 7109c6f1ddSLingrui98 7209c6f1ddSLingrui98// record the situation in which fallThruAddr falls into 7309c6f1ddSLingrui98// the middle of an RVI inst 7409c6f1ddSLingrui98class LastHalfInfo(implicit p: Parameters) extends XSBundle { 7509c6f1ddSLingrui98 val valid = Bool() 7609c6f1ddSLingrui98 val middlePC = UInt(VAddrBits.W) 7709c6f1ddSLingrui98 def matchThisBlock(startAddr: UInt) = valid && middlePC === startAddr 7809c6f1ddSLingrui98} 7909c6f1ddSLingrui98 8009c6f1ddSLingrui98class IfuToPreDecode(implicit p: Parameters) extends XSBundle { 8109c6f1ddSLingrui98 val data = if(HasCExtension) Vec(PredictWidth + 1, UInt(16.W)) else Vec(PredictWidth, UInt(32.W)) 8272951335SLi Qianruo val frontendTrigger = new FrontendTdataDistributeIO 8372951335SLi Qianruo val csrTriggerEnable = Vec(4, Bool()) 842a3050c2SJay val pc = Vec(PredictWidth, UInt(VAddrBits.W)) 8509c6f1ddSLingrui98} 8609c6f1ddSLingrui98 872a3050c2SJay 882a3050c2SJayclass IfuToPredChecker(implicit p: Parameters) extends XSBundle { 892a3050c2SJay val ftqOffset = Valid(UInt(log2Ceil(PredictWidth).W)) 902a3050c2SJay val jumpOffset = Vec(PredictWidth, UInt(XLEN.W)) 912a3050c2SJay val target = UInt(VAddrBits.W) 922a3050c2SJay val instrRange = Vec(PredictWidth, Bool()) 932a3050c2SJay val instrValid = Vec(PredictWidth, Bool()) 942a3050c2SJay val pds = Vec(PredictWidth, new PreDecodeInfo) 952a3050c2SJay val pc = Vec(PredictWidth, UInt(VAddrBits.W)) 962a3050c2SJay} 972a3050c2SJay 982a3050c2SJayclass NewIFU(implicit p: Parameters) extends XSModule 992a3050c2SJay with HasICacheParameters 1002a3050c2SJay with HasIFUConst 1012a3050c2SJay with HasPdConst 102167bcd01SJay with HasCircularQueuePtrHelper 1032a3050c2SJay with HasPerfEvents 10409c6f1ddSLingrui98{ 10509c6f1ddSLingrui98 val io = IO(new NewIFUIO) 10609c6f1ddSLingrui98 val (toFtq, fromFtq) = (io.ftqInter.toFtq, io.ftqInter.fromFtq) 107c5c5edaeSJenius val fromICache = io.icacheInter.resp 1080be662e4SJay val (toUncache, fromUncache) = (io.uncacheInter.toUncache , io.uncacheInter.fromUncache) 10909c6f1ddSLingrui98 11009c6f1ddSLingrui98 def isCrossLineReq(start: UInt, end: UInt): Bool = start(blockOffBits) ^ end(blockOffBits) 11109c6f1ddSLingrui98 11234a88126SJinYue def isLastInCacheline(addr: UInt): Bool = addr(blockOffBits - 1, 1) === 0.U 11309c6f1ddSLingrui98 1141d8f4dcbSJay class TlbExept(implicit p: Parameters) extends XSBundle{ 1151d8f4dcbSJay val pageFault = Bool() 1161d8f4dcbSJay val accessFault = Bool() 1171d8f4dcbSJay val mmio = Bool() 118b005f7c6SJay } 11909c6f1ddSLingrui98 120dc270d3bSJenius val preDecoders = Seq.fill(4){ Module(new PreDecode) } 121dc270d3bSJenius 1222a3050c2SJay val predChecker = Module(new PredChecker) 1232a3050c2SJay val frontendTrigger = Module(new FrontendTrigger) 1245995c9e7SJenius val (checkerIn, checkerOutStage1, checkerOutStage2) = (predChecker.io.in, predChecker.io.out.stage1Out,predChecker.io.out.stage2Out) 1251d8f4dcbSJay 126c3b763d0SYinan Xu io.iTLBInter.req_kill := false.B 127ee175d78SJay io.iTLBInter.resp.ready := true.B 128ee175d78SJay 12958dbdfc2SJay /** 13058dbdfc2SJay ****************************************************************************** 13158dbdfc2SJay * IFU Stage 0 13258dbdfc2SJay * - send cacheline fetch request to ICacheMainPipe 13358dbdfc2SJay ****************************************************************************** 13458dbdfc2SJay */ 13509c6f1ddSLingrui98 13609c6f1ddSLingrui98 val f0_valid = fromFtq.req.valid 13709c6f1ddSLingrui98 val f0_ftq_req = fromFtq.req.bits 1386ce52296SJinYue val f0_doubleLine = fromFtq.req.bits.crossCacheline 13934a88126SJinYue val f0_vSetIdx = VecInit(get_idx((f0_ftq_req.startAddr)), get_idx(f0_ftq_req.nextlineStart)) 14009c6f1ddSLingrui98 val f0_fire = fromFtq.req.fire() 14109c6f1ddSLingrui98 14209c6f1ddSLingrui98 val f0_flush, f1_flush, f2_flush, f3_flush = WireInit(false.B) 14309c6f1ddSLingrui98 val from_bpu_f0_flush, from_bpu_f1_flush, from_bpu_f2_flush, from_bpu_f3_flush = WireInit(false.B) 14409c6f1ddSLingrui98 145cb4f77ceSLingrui98 from_bpu_f0_flush := fromFtq.flushFromBpu.shouldFlushByStage2(f0_ftq_req.ftqIdx) || 146cb4f77ceSLingrui98 fromFtq.flushFromBpu.shouldFlushByStage3(f0_ftq_req.ftqIdx) 14709c6f1ddSLingrui98 1482a3050c2SJay val wb_redirect , mmio_redirect, backend_redirect= WireInit(false.B) 1492a3050c2SJay val f3_wb_not_flush = WireInit(false.B) 1502a3050c2SJay 1512a3050c2SJay backend_redirect := fromFtq.redirect.valid 1522a3050c2SJay f3_flush := backend_redirect || (wb_redirect && !f3_wb_not_flush) 1532a3050c2SJay f2_flush := backend_redirect || mmio_redirect || wb_redirect 15409c6f1ddSLingrui98 f1_flush := f2_flush || from_bpu_f1_flush 15509c6f1ddSLingrui98 f0_flush := f1_flush || from_bpu_f0_flush 15609c6f1ddSLingrui98 15709c6f1ddSLingrui98 val f1_ready, f2_ready, f3_ready = WireInit(false.B) 15809c6f1ddSLingrui98 15950780602SJenius fromFtq.req.ready := f1_ready && io.icacheInter.icacheReady 16009c6f1ddSLingrui98 16158dbdfc2SJay /** <PERF> f0 fetch bubble */ 162f7c29b0aSJinYue 16300240ba6SJay XSPerfAccumulate("fetch_bubble_ftq_not_valid", !fromFtq.req.valid && fromFtq.req.ready ) 164c5c5edaeSJenius // XSPerfAccumulate("fetch_bubble_pipe_stall", f0_valid && toICache(0).ready && toICache(1).ready && !f1_ready ) 165c5c5edaeSJenius // XSPerfAccumulate("fetch_bubble_icache_0_busy", f0_valid && !toICache(0).ready ) 166c5c5edaeSJenius // XSPerfAccumulate("fetch_bubble_icache_1_busy", f0_valid && !toICache(1).ready ) 16700240ba6SJay XSPerfAccumulate("fetch_flush_backend_redirect", backend_redirect ) 16800240ba6SJay XSPerfAccumulate("fetch_flush_wb_redirect", wb_redirect ) 16900240ba6SJay XSPerfAccumulate("fetch_flush_bpu_f1_flush", from_bpu_f1_flush ) 17000240ba6SJay XSPerfAccumulate("fetch_flush_bpu_f0_flush", from_bpu_f0_flush ) 17158dbdfc2SJay 17258dbdfc2SJay 17358dbdfc2SJay /** 17458dbdfc2SJay ****************************************************************************** 17558dbdfc2SJay * IFU Stage 1 17658dbdfc2SJay * - calculate pc/half_pc/cut_ptr for every instruction 17758dbdfc2SJay ****************************************************************************** 17858dbdfc2SJay */ 17909c6f1ddSLingrui98 18009c6f1ddSLingrui98 val f1_valid = RegInit(false.B) 181005e809bSJiuyang Liu val f1_ftq_req = RegEnable(f0_ftq_req, f0_fire) 182005e809bSJiuyang Liu // val f1_situation = RegEnable(f0_situation, f0_fire) 183005e809bSJiuyang Liu val f1_doubleLine = RegEnable(f0_doubleLine, f0_fire) 184005e809bSJiuyang Liu val f1_vSetIdx = RegEnable(f0_vSetIdx, f0_fire) 185625ecd17SJenius val f1_fire = f1_valid && f2_ready 18609c6f1ddSLingrui98 187625ecd17SJenius f1_ready := f1_fire || !f1_valid 18809c6f1ddSLingrui98 1890d756c48SJinYue from_bpu_f1_flush := fromFtq.flushFromBpu.shouldFlushByStage3(f1_ftq_req.ftqIdx) && f1_valid 190cb4f77ceSLingrui98 // from_bpu_f1_flush := false.B 19109c6f1ddSLingrui98 19209c6f1ddSLingrui98 when(f1_flush) {f1_valid := false.B} 19309c6f1ddSLingrui98 .elsewhen(f0_fire && !f0_flush) {f1_valid := true.B} 19409c6f1ddSLingrui98 .elsewhen(f1_fire) {f1_valid := false.B} 19509c6f1ddSLingrui98 1962a3050c2SJay val f1_pc = VecInit((0 until PredictWidth).map(i => f1_ftq_req.startAddr + (i * 2).U)) 1972a3050c2SJay val f1_half_snpc = VecInit((0 until PredictWidth).map(i => f1_ftq_req.startAddr + ((i+2) * 2).U)) 1982a3050c2SJay val f1_cut_ptr = if(HasCExtension) VecInit((0 until PredictWidth + 1).map(i => Cat(0.U(1.W), f1_ftq_req.startAddr(blockOffBits-1, 1)) + i.U )) 1992a3050c2SJay else VecInit((0 until PredictWidth).map(i => Cat(0.U(1.W), f1_ftq_req.startAddr(blockOffBits-1, 2)) + i.U )) 20009c6f1ddSLingrui98 20158dbdfc2SJay /** 20258dbdfc2SJay ****************************************************************************** 20358dbdfc2SJay * IFU Stage 2 20458dbdfc2SJay * - icache response data (latched for pipeline stop) 20558dbdfc2SJay * - generate exceprion bits for every instruciton (page fault/access fault/mmio) 20658dbdfc2SJay * - generate predicted instruction range (1 means this instruciton is in this fetch packet) 20758dbdfc2SJay * - cut data from cachlines to packet instruction code 20858dbdfc2SJay * - instruction predecode and RVC expand 20958dbdfc2SJay ****************************************************************************** 21058dbdfc2SJay */ 21158dbdfc2SJay 2121d8f4dcbSJay val icacheRespAllValid = WireInit(false.B) 21309c6f1ddSLingrui98 21409c6f1ddSLingrui98 val f2_valid = RegInit(false.B) 215005e809bSJiuyang Liu val f2_ftq_req = RegEnable(f1_ftq_req, f1_fire) 216005e809bSJiuyang Liu // val f2_situation = RegEnable(f1_situation, f1_fire) 217005e809bSJiuyang Liu val f2_doubleLine = RegEnable(f1_doubleLine, f1_fire) 218005e809bSJiuyang Liu val f2_vSetIdx = RegEnable(f1_vSetIdx, f1_fire) 219625ecd17SJenius val f2_fire = f2_valid && f3_ready && icacheRespAllValid 2201d8f4dcbSJay 221625ecd17SJenius f2_ready := f2_fire || !f2_valid 2221d8f4dcbSJay //TODO: addr compare may be timing critical 22334a88126SJinYue val f2_icache_all_resp_wire = fromICache(0).valid && (fromICache(0).bits.vaddr === f2_ftq_req.startAddr) && ((fromICache(1).valid && (fromICache(1).bits.vaddr === f2_ftq_req.nextlineStart)) || !f2_doubleLine) 2241d8f4dcbSJay val f2_icache_all_resp_reg = RegInit(false.B) 2251d8f4dcbSJay 2261d8f4dcbSJay icacheRespAllValid := f2_icache_all_resp_reg || f2_icache_all_resp_wire 2271d8f4dcbSJay 2281d8f4dcbSJay io.icacheStop := !f3_ready 2291d8f4dcbSJay 2301d8f4dcbSJay when(f2_flush) {f2_icache_all_resp_reg := false.B} 2311d8f4dcbSJay .elsewhen(f2_valid && f2_icache_all_resp_wire && !f3_ready) {f2_icache_all_resp_reg := true.B} 2321d8f4dcbSJay .elsewhen(f2_fire && f2_icache_all_resp_reg) {f2_icache_all_resp_reg := false.B} 23309c6f1ddSLingrui98 23409c6f1ddSLingrui98 when(f2_flush) {f2_valid := false.B} 23509c6f1ddSLingrui98 .elsewhen(f1_fire && !f1_flush) {f2_valid := true.B } 23609c6f1ddSLingrui98 .elsewhen(f2_fire) {f2_valid := false.B} 23709c6f1ddSLingrui98 2380bca1ccbSJinYue // val f2_cache_response_data = ResultHoldBypass(valid = f2_icache_all_resp_wire, data = VecInit(fromICache.map(_.bits.readData))) 239dc270d3bSJenius val f2_cache_response_reg_data = VecInit(fromICache.map(_.bits.registerData)) 240dc270d3bSJenius val f2_cache_response_sram_data = VecInit(fromICache.map(_.bits.sramData)) 241dc270d3bSJenius val f2_cache_response_select = VecInit(fromICache.map(_.bits.select)) 2420bca1ccbSJinYue 24309c6f1ddSLingrui98 2441d8f4dcbSJay val f2_except_pf = VecInit((0 until PortNumber).map(i => fromICache(i).bits.tlbExcp.pageFault)) 2451d8f4dcbSJay val f2_except_af = VecInit((0 until PortNumber).map(i => fromICache(i).bits.tlbExcp.accessFault)) 246c0b2b8e9Srvcoresjw val f2_mmio = fromICache(0).bits.tlbExcp.mmio && !fromICache(0).bits.tlbExcp.accessFault && 247c0b2b8e9Srvcoresjw !fromICache(0).bits.tlbExcp.pageFault 2480be662e4SJay 249005e809bSJiuyang Liu val f2_pc = RegEnable(f1_pc, f1_fire) 250005e809bSJiuyang Liu val f2_half_snpc = RegEnable(f1_half_snpc, f1_fire) 251005e809bSJiuyang Liu val f2_cut_ptr = RegEnable(f1_cut_ptr, f1_fire) 252a37fbf10SJay 253005e809bSJiuyang Liu val f2_resend_vaddr = RegEnable(f1_ftq_req.startAddr + 2.U, f1_fire) 2542a3050c2SJay 2552a3050c2SJay def isNextLine(pc: UInt, startAddr: UInt) = { 2562a3050c2SJay startAddr(blockOffBits) ^ pc(blockOffBits) 257b6982e83SLemover } 25809c6f1ddSLingrui98 2592a3050c2SJay def isLastInLine(pc: UInt) = { 2602a3050c2SJay pc(blockOffBits - 1, 0) === "b111110".U 26109c6f1ddSLingrui98 } 26209c6f1ddSLingrui98 2632a3050c2SJay val f2_foldpc = VecInit(f2_pc.map(i => XORFold(i(VAddrBits-1,1), MemPredPCWidth))) 2642a3050c2SJay val f2_jump_range = Fill(PredictWidth, !f2_ftq_req.ftqOffset.valid) | Fill(PredictWidth, 1.U(1.W)) >> ~f2_ftq_req.ftqOffset.bits 2651d011975SJinYue val f2_ftr_range = Fill(PredictWidth, f2_ftq_req.ftqOffset.valid) | Fill(PredictWidth, 1.U(1.W)) >> ~getBasicBlockIdx(f2_ftq_req.nextStartAddr, f2_ftq_req.startAddr) 2662a3050c2SJay val f2_instr_range = f2_jump_range & f2_ftr_range 2672a3050c2SJay val f2_pf_vec = VecInit((0 until PredictWidth).map(i => (!isNextLine(f2_pc(i), f2_ftq_req.startAddr) && f2_except_pf(0) || isNextLine(f2_pc(i), f2_ftq_req.startAddr) && f2_doubleLine && f2_except_pf(1)))) 2682a3050c2SJay val f2_af_vec = VecInit((0 until PredictWidth).map(i => (!isNextLine(f2_pc(i), f2_ftq_req.startAddr) && f2_except_af(0) || isNextLine(f2_pc(i), f2_ftq_req.startAddr) && f2_doubleLine && f2_except_af(1)))) 26909c6f1ddSLingrui98 2701d8f4dcbSJay val f2_paddrs = VecInit((0 until PortNumber).map(i => fromICache(i).bits.paddr)) 2711d8f4dcbSJay val f2_perf_info = io.icachePerfInfo 27209c6f1ddSLingrui98 2732a3050c2SJay def cut(cacheline: UInt, cutPtr: Vec[UInt]) : Vec[UInt] ={ 274d558bd61SJenius require(HasCExtension) 275d558bd61SJenius // if(HasCExtension){ 276d558bd61SJenius val partCacheline = cacheline((blockBytes * 8 * 2 * 3) / 4 - 1, 0) 27709c6f1ddSLingrui98 val result = Wire(Vec(PredictWidth + 1, UInt(16.W))) 278d558bd61SJenius val dataVec = cacheline.asTypeOf(Vec(blockBytes * 3 /4, UInt(16.W))) //47 16-bit data vector 27909c6f1ddSLingrui98 (0 until PredictWidth + 1).foreach( i => 280d558bd61SJenius result(i) := dataVec(cutPtr(i)) //the max ptr is 3*blockBytes/4-1 28109c6f1ddSLingrui98 ) 28209c6f1ddSLingrui98 result 283d558bd61SJenius // } else { 284d558bd61SJenius // val result = Wire(Vec(PredictWidth, UInt(32.W)) ) 285d558bd61SJenius // val dataVec = cacheline.asTypeOf(Vec(blockBytes * 2/ 4, UInt(32.W))) 286d558bd61SJenius // (0 until PredictWidth).foreach( i => 287d558bd61SJenius // result(i) := dataVec(cutPtr(i)) 288d558bd61SJenius // ) 289d558bd61SJenius // result 290d558bd61SJenius // } 29109c6f1ddSLingrui98 } 29209c6f1ddSLingrui98 293dc270d3bSJenius val f2_data_2_cacheline = Wire(Vec(4, UInt((2 * blockBits).W))) 294dc270d3bSJenius f2_data_2_cacheline(0) := Cat(f2_cache_response_reg_data(1) , f2_cache_response_reg_data(0)) 295dc270d3bSJenius f2_data_2_cacheline(1) := Cat(f2_cache_response_reg_data(1) , f2_cache_response_sram_data(0)) 296dc270d3bSJenius f2_data_2_cacheline(2) := Cat(f2_cache_response_sram_data(1) , f2_cache_response_reg_data(0)) 297dc270d3bSJenius f2_data_2_cacheline(3) := Cat(f2_cache_response_sram_data(1) , f2_cache_response_sram_data(0)) 298dc270d3bSJenius 299dc270d3bSJenius val f2_cut_data = VecInit(f2_data_2_cacheline.map(data => cut( data, f2_cut_ptr ))) 300dc270d3bSJenius 301dc270d3bSJenius val f2_predecod_ptr = Wire(UInt(2.W)) 302dc270d3bSJenius f2_predecod_ptr := Cat(f2_cache_response_select(1),f2_cache_response_select(0)) 30309c6f1ddSLingrui98 30458dbdfc2SJay /** predecode (include RVC expander) */ 305dc270d3bSJenius // preDecoderRegIn.data := f2_reg_cut_data 306dc270d3bSJenius // preDecoderRegInIn.frontendTrigger := io.frontendTrigger 307dc270d3bSJenius // preDecoderRegInIn.csrTriggerEnable := io.csrTriggerEnable 308dc270d3bSJenius // preDecoderRegIn.pc := f2_pc 309dc270d3bSJenius 310dc270d3bSJenius val preDecoderOut = Mux1H(UIntToOH(f2_predecod_ptr), preDecoders.map(_.io.out)) 311dc270d3bSJenius for(i <- 0 until 4){ 312dc270d3bSJenius val preDecoderIn = preDecoders(i).io.in 313dc270d3bSJenius preDecoderIn.data := f2_cut_data(i) 3142a3050c2SJay preDecoderIn.frontendTrigger := io.frontendTrigger 3152a3050c2SJay preDecoderIn.csrTriggerEnable := io.csrTriggerEnable 3162a3050c2SJay preDecoderIn.pc := f2_pc 317dc270d3bSJenius } 31809c6f1ddSLingrui98 31948a62719SJenius //val f2_expd_instr = preDecoderOut.expInstr 32048a62719SJenius val f2_instr = preDecoderOut.instr 3212a3050c2SJay val f2_pd = preDecoderOut.pd 3222a3050c2SJay val f2_jump_offset = preDecoderOut.jumpOffset 3232a3050c2SJay val f2_hasHalfValid = preDecoderOut.hasHalfValid 3242a3050c2SJay val f2_crossPageFault = VecInit((0 until PredictWidth).map(i => isLastInLine(f2_pc(i)) && !f2_except_pf(0) && f2_doubleLine && f2_except_pf(1) && !f2_pd(i).isRVC )) 32509c6f1ddSLingrui98 32600240ba6SJay XSPerfAccumulate("fetch_bubble_icache_not_resp", f2_valid && !icacheRespAllValid ) 32700240ba6SJay 32809c6f1ddSLingrui98 32958dbdfc2SJay /** 33058dbdfc2SJay ****************************************************************************** 33158dbdfc2SJay * IFU Stage 3 33258dbdfc2SJay * - handle MMIO instruciton 33358dbdfc2SJay * -send request to Uncache fetch Unit 33458dbdfc2SJay * -every packet include 1 MMIO instruction 33558dbdfc2SJay * -MMIO instructions will stop fetch pipeline until commiting from RoB 33658dbdfc2SJay * -flush to snpc (send ifu_redirect to Ftq) 33758dbdfc2SJay * - Ibuffer enqueue 33858dbdfc2SJay * - check predict result in Frontend (jalFault/retFault/notCFIFault/invalidTakenFault/targetFault) 33958dbdfc2SJay * - handle last half RVI instruction 34058dbdfc2SJay ****************************************************************************** 34158dbdfc2SJay */ 34258dbdfc2SJay 34309c6f1ddSLingrui98 val f3_valid = RegInit(false.B) 344005e809bSJiuyang Liu val f3_ftq_req = RegEnable(f2_ftq_req, f2_fire) 345005e809bSJiuyang Liu // val f3_situation = RegEnable(f2_situation, f2_fire) 346005e809bSJiuyang Liu val f3_doubleLine = RegEnable(f2_doubleLine, f2_fire) 3471d8f4dcbSJay val f3_fire = io.toIbuffer.fire() 3481d8f4dcbSJay 349625ecd17SJenius f3_ready := f3_fire || !f3_valid 35009c6f1ddSLingrui98 351dc270d3bSJenius val f3_cut_data = RegEnable(next = f2_cut_data(f2_predecod_ptr), enable=f2_fire) 3521d8f4dcbSJay 353005e809bSJiuyang Liu val f3_except_pf = RegEnable(f2_except_pf, f2_fire) 354005e809bSJiuyang Liu val f3_except_af = RegEnable(f2_except_af, f2_fire) 355005e809bSJiuyang Liu val f3_mmio = RegEnable(f2_mmio , f2_fire) 35609c6f1ddSLingrui98 35748a62719SJenius //val f3_expd_instr = RegEnable(next = f2_expd_instr, enable = f2_fire) 35848a62719SJenius val f3_instr = RegEnable(next = f2_instr, enable = f2_fire) 35948a62719SJenius val f3_expd_instr = VecInit((0 until PredictWidth).map{ i => 36048a62719SJenius val expander = Module(new RVCExpander) 36148a62719SJenius expander.io.in := f3_instr(i) 36248a62719SJenius expander.io.out.bits 36348a62719SJenius }) 36448a62719SJenius 36548a62719SJenius val f3_pd = RegEnable(next = f2_pd, enable = f2_fire) 36648a62719SJenius val f3_jump_offset = RegEnable(next = f2_jump_offset, enable = f2_fire) 36748a62719SJenius val f3_af_vec = RegEnable(next = f2_af_vec, enable = f2_fire) 36848a62719SJenius val f3_pf_vec = RegEnable(next = f2_pf_vec , enable = f2_fire) 36948a62719SJenius val f3_pc = RegEnable(next = f2_pc, enable = f2_fire) 37048a62719SJenius val f3_half_snpc = RegEnable(next = f2_half_snpc, enable = f2_fire) 37148a62719SJenius val f3_instr_range = RegEnable(next = f2_instr_range, enable = f2_fire) 37248a62719SJenius val f3_foldpc = RegEnable(next = f2_foldpc, enable = f2_fire) 37348a62719SJenius val f3_crossPageFault = RegEnable(next = f2_crossPageFault, enable = f2_fire) 37448a62719SJenius val f3_hasHalfValid = RegEnable(next = f2_hasHalfValid, enable = f2_fire) 37509c6f1ddSLingrui98 val f3_except = VecInit((0 until 2).map{i => f3_except_pf(i) || f3_except_af(i)}) 37609c6f1ddSLingrui98 val f3_has_except = f3_valid && (f3_except_af.reduce(_||_) || f3_except_pf.reduce(_||_)) 377005e809bSJiuyang Liu val f3_pAddrs = RegEnable(f2_paddrs, f2_fire) 378005e809bSJiuyang Liu val f3_resend_vaddr = RegEnable(f2_resend_vaddr, f2_fire) 379ee175d78SJay 3801d011975SJinYue when(f3_valid && !f3_ftq_req.ftqOffset.valid){ 3811d011975SJinYue assert(f3_ftq_req.startAddr + 32.U >= f3_ftq_req.nextStartAddr , "More tha 32 Bytes fetch is not allowed!") 3821d011975SJinYue } 383a1351e5dSJay 3842a3050c2SJay /*** MMIO State Machine***/ 385ee175d78SJay val f3_mmio_data = Reg(Vec(2, UInt(16.W))) 386ee175d78SJay val mmio_is_RVC = RegInit(false.B) 387ee175d78SJay val mmio_resend_addr =RegInit(0.U(PAddrBits.W)) 388ee175d78SJay val mmio_resend_af = RegInit(false.B) 389c3b2d83aSJay val mmio_resend_pf = RegInit(false.B) 390c3b2d83aSJay 391a37fbf10SJay 392ee175d78SJay val m_idle :: m_sendReq :: m_waitResp :: m_sendTLB :: m_tlbResp :: m_sendPMP :: m_resendReq :: m_waitResendResp :: m_waitCommit :: m_commited :: Nil = Enum(10) 393ee175d78SJay val mmio_state = RegInit(m_idle) 394a37fbf10SJay 3959bae7d6eSJay val f3_req_is_mmio = f3_mmio && f3_valid 3962a3050c2SJay val mmio_commit = VecInit(io.rob_commits.map{commit => commit.valid && commit.bits.ftqIdx === f3_ftq_req.ftqIdx && commit.bits.ftqOffset === 0.U}).asUInt.orR 397ee175d78SJay val f3_mmio_req_commit = f3_req_is_mmio && mmio_state === m_commited 398a37fbf10SJay 399ee175d78SJay val f3_mmio_to_commit = f3_req_is_mmio && mmio_state === m_waitCommit 400a37fbf10SJay val f3_mmio_to_commit_next = RegNext(f3_mmio_to_commit) 401a37fbf10SJay val f3_mmio_can_go = f3_mmio_to_commit && !f3_mmio_to_commit_next 402a37fbf10SJay 40356788a33SJinYue val fromFtqRedirectReg = RegNext(fromFtq.redirect) 40456788a33SJinYue val f3_ftq_flush_self = fromFtqRedirectReg.valid && RedirectLevel.flushItself(fromFtqRedirectReg.bits.level) 40556788a33SJinYue val f3_ftq_flush_by_older = fromFtqRedirectReg.valid && isBefore(fromFtqRedirectReg.bits.ftqIdx, f3_ftq_req.ftqIdx) 4069bae7d6eSJay 40756788a33SJinYue val f3_need_not_flush = f3_req_is_mmio && fromFtqRedirectReg.valid && !f3_ftq_flush_self && !f3_ftq_flush_by_older 4089bae7d6eSJay 4099bae7d6eSJay when(f3_flush && !f3_need_not_flush) {f3_valid := false.B} 410a37fbf10SJay .elsewhen(f2_fire && !f2_flush ) {f3_valid := true.B } 411a37fbf10SJay .elsewhen(io.toIbuffer.fire() && !f3_req_is_mmio) {f3_valid := false.B} 412a37fbf10SJay .elsewhen{f3_req_is_mmio && f3_mmio_req_commit} {f3_valid := false.B} 413a37fbf10SJay 414a37fbf10SJay val f3_mmio_use_seq_pc = RegInit(false.B) 415a37fbf10SJay 41656788a33SJinYue val (redirect_ftqIdx, redirect_ftqOffset) = (fromFtqRedirectReg.bits.ftqIdx,fromFtqRedirectReg.bits.ftqOffset) 41756788a33SJinYue val redirect_mmio_req = fromFtqRedirectReg.valid && redirect_ftqIdx === f3_ftq_req.ftqIdx && redirect_ftqOffset === 0.U 418a37fbf10SJay 419a37fbf10SJay when(RegNext(f2_fire && !f2_flush) && f3_req_is_mmio) { f3_mmio_use_seq_pc := true.B } 420a37fbf10SJay .elsewhen(redirect_mmio_req) { f3_mmio_use_seq_pc := false.B } 421a37fbf10SJay 422a37fbf10SJay f3_ready := Mux(f3_req_is_mmio, io.toIbuffer.ready && f3_mmio_req_commit || !f3_valid , io.toIbuffer.ready || !f3_valid) 423a37fbf10SJay 424a37fbf10SJay 425a37fbf10SJay switch(mmio_state){ 426ee175d78SJay is(m_idle){ 4279bae7d6eSJay when(f3_req_is_mmio){ 428ee175d78SJay mmio_state := m_sendReq 429a37fbf10SJay } 430a37fbf10SJay } 431a37fbf10SJay 432ee175d78SJay is(m_sendReq){ 433ee175d78SJay mmio_state := Mux(toUncache.fire(), m_waitResp, m_sendReq ) 434a37fbf10SJay } 435a37fbf10SJay 436ee175d78SJay is(m_waitResp){ 437a37fbf10SJay when(fromUncache.fire()){ 438a37fbf10SJay val isRVC = fromUncache.bits.data(1,0) =/= 3.U 439ee175d78SJay val needResend = !isRVC && f3_pAddrs(0)(2,1) === 3.U 440ee175d78SJay mmio_state := Mux(needResend, m_sendTLB , m_waitCommit) 441ee175d78SJay 442ee175d78SJay mmio_is_RVC := isRVC 443ee175d78SJay f3_mmio_data(0) := fromUncache.bits.data(15,0) 444ee175d78SJay f3_mmio_data(1) := fromUncache.bits.data(31,16) 445a37fbf10SJay } 446a37fbf10SJay } 447a37fbf10SJay 448ee175d78SJay is(m_sendTLB){ 449c3b2d83aSJay when( io.iTLBInter.req.valid && !io.iTLBInter.resp.bits.miss ){ 450ee175d78SJay mmio_state := m_tlbResp 451a37fbf10SJay } 452c3b2d83aSJay } 453a37fbf10SJay 454ee175d78SJay is(m_tlbResp){ 45503efd994Shappy-lx val tlbExept = io.iTLBInter.resp.bits.excp(0).pf.instr || 45603efd994Shappy-lx io.iTLBInter.resp.bits.excp(0).af.instr 457c3b2d83aSJay mmio_state := Mux(tlbExept,m_waitCommit,m_sendPMP) 45803efd994Shappy-lx mmio_resend_addr := io.iTLBInter.resp.bits.paddr(0) 459920ca00eSJay mmio_resend_af := mmio_resend_af || io.iTLBInter.resp.bits.excp(0).af.instr 460920ca00eSJay mmio_resend_pf := mmio_resend_pf || io.iTLBInter.resp.bits.excp(0).pf.instr 461ee175d78SJay } 462ee175d78SJay 463ee175d78SJay is(m_sendPMP){ 464c3b2d83aSJay val pmpExcpAF = io.pmp.resp.instr || !io.pmp.resp.mmio 465ee175d78SJay mmio_state := Mux(pmpExcpAF, m_waitCommit , m_resendReq) 466ee175d78SJay mmio_resend_af := pmpExcpAF 467ee175d78SJay } 468ee175d78SJay 469ee175d78SJay is(m_resendReq){ 470ee175d78SJay mmio_state := Mux(toUncache.fire(), m_waitResendResp, m_resendReq ) 471ee175d78SJay } 472ee175d78SJay 473ee175d78SJay is(m_waitResendResp){ 474a37fbf10SJay when(fromUncache.fire()){ 475ee175d78SJay mmio_state := m_waitCommit 476ee175d78SJay f3_mmio_data(1) := fromUncache.bits.data(15,0) 477a37fbf10SJay } 478a37fbf10SJay } 479a37fbf10SJay 480ee175d78SJay is(m_waitCommit){ 4812a3050c2SJay when(mmio_commit){ 482ee175d78SJay mmio_state := m_commited 483a37fbf10SJay } 484a37fbf10SJay } 4852a3050c2SJay 486ee175d78SJay //normal mmio instruction 487ee175d78SJay is(m_commited){ 488ee175d78SJay mmio_state := m_idle 489ee175d78SJay mmio_is_RVC := false.B 490ee175d78SJay mmio_resend_addr := 0.U 4912a3050c2SJay } 492a37fbf10SJay } 493a37fbf10SJay 494ee175d78SJay //exception or flush by older branch prediction 495167bcd01SJay when(f3_ftq_flush_self || f3_ftq_flush_by_older) { 496ee175d78SJay mmio_state := m_idle 497ee175d78SJay mmio_is_RVC := false.B 498ee175d78SJay mmio_resend_addr := 0.U 499ee175d78SJay mmio_resend_af := false.B 500ee175d78SJay f3_mmio_data.map(_ := 0.U) 5019bae7d6eSJay } 5029bae7d6eSJay 503ee175d78SJay toUncache.valid := ((mmio_state === m_sendReq) || (mmio_state === m_resendReq)) && f3_req_is_mmio 504ee175d78SJay toUncache.bits.addr := Mux((mmio_state === m_resendReq), mmio_resend_addr, f3_pAddrs(0)) 505a37fbf10SJay fromUncache.ready := true.B 506a37fbf10SJay 507ee175d78SJay io.iTLBInter.req.valid := (mmio_state === m_sendTLB) && f3_req_is_mmio 508ee175d78SJay io.iTLBInter.req.bits.size := 3.U 509ee175d78SJay io.iTLBInter.req.bits.vaddr := f3_resend_vaddr 510ee175d78SJay io.iTLBInter.req.bits.debug.pc := f3_resend_vaddr 511ee175d78SJay 512f1fe8698SLemover io.iTLBInter.req.bits.kill := false.B // IFU use itlb for mmio, doesn't need sync, set it to false 513ee175d78SJay io.iTLBInter.req.bits.cmd := TlbCmd.exec 514f1fe8698SLemover io.iTLBInter.req.bits.debug.robIdx := DontCare 515ee175d78SJay io.iTLBInter.req.bits.debug.isFirstIssue := DontCare 516ee175d78SJay 517ee175d78SJay io.pmp.req.valid := (mmio_state === m_sendPMP) && f3_req_is_mmio 518ee175d78SJay io.pmp.req.bits.addr := mmio_resend_addr 519ee175d78SJay io.pmp.req.bits.size := 3.U 520ee175d78SJay io.pmp.req.bits.cmd := TlbCmd.exec 521f7c29b0aSJinYue 5222a3050c2SJay val f3_lastHalf = RegInit(0.U.asTypeOf(new LastHalfInfo)) 52309c6f1ddSLingrui98 52409c6f1ddSLingrui98 val f3_predecode_range = VecInit(preDecoderOut.pd.map(inst => inst.valid)).asUInt 5250be662e4SJay val f3_mmio_range = VecInit((0 until PredictWidth).map(i => if(i ==0) true.B else false.B)) 5262a3050c2SJay val f3_instr_valid = Wire(Vec(PredictWidth, Bool())) 52709c6f1ddSLingrui98 5282a3050c2SJay /*** prediction result check ***/ 5292a3050c2SJay checkerIn.ftqOffset := f3_ftq_req.ftqOffset 5302a3050c2SJay checkerIn.jumpOffset := f3_jump_offset 5316ce52296SJinYue checkerIn.target := f3_ftq_req.nextStartAddr 5322a3050c2SJay checkerIn.instrRange := f3_instr_range.asTypeOf(Vec(PredictWidth, Bool())) 5332a3050c2SJay checkerIn.instrValid := f3_instr_valid.asTypeOf(Vec(PredictWidth, Bool())) 5342a3050c2SJay checkerIn.pds := f3_pd 5352a3050c2SJay checkerIn.pc := f3_pc 5362a3050c2SJay 53758dbdfc2SJay /*** handle half RVI in the last 2 Bytes ***/ 5382a3050c2SJay 5392a3050c2SJay def hasLastHalf(idx: UInt) = { 5405995c9e7SJenius //!f3_pd(idx).isRVC && checkerOutStage1.fixedRange(idx) && f3_instr_valid(idx) && !checkerOutStage1.fixedTaken(idx) && !checkerOutStage2.fixedMissPred(idx) && ! f3_req_is_mmio 5415995c9e7SJenius !f3_pd(idx).isRVC && checkerOutStage1.fixedRange(idx) && f3_instr_valid(idx) && !checkerOutStage1.fixedTaken(idx) && ! f3_req_is_mmio 5422a3050c2SJay } 5432a3050c2SJay 5445995c9e7SJenius val f3_last_validIdx = ~ParallelPriorityEncoder(checkerOutStage1.fixedRange.reverse) 5452a3050c2SJay 5462a3050c2SJay val f3_hasLastHalf = hasLastHalf((PredictWidth - 1).U) 5472a3050c2SJay val f3_false_lastHalf = hasLastHalf(f3_last_validIdx) 5482a3050c2SJay val f3_false_snpc = f3_half_snpc(f3_last_validIdx) 5492a3050c2SJay 5502a3050c2SJay val f3_lastHalf_mask = VecInit((0 until PredictWidth).map( i => if(i ==0) false.B else true.B )).asUInt() 5513f785aa3SJenius val f3_lastHalf_disable = RegInit(false.B) 5522a3050c2SJay 553804985a5SJenius when(f3_flush || (f3_fire && f3_lastHalf_disable)){ 554804985a5SJenius f3_lastHalf_disable := false.B 555804985a5SJenius } 556804985a5SJenius 5572a3050c2SJay when (f3_flush) { 5582a3050c2SJay f3_lastHalf.valid := false.B 5592a3050c2SJay }.elsewhen (f3_fire) { 5603f785aa3SJenius f3_lastHalf.valid := f3_hasLastHalf && !f3_lastHalf_disable 5616ce52296SJinYue f3_lastHalf.middlePC := f3_ftq_req.nextStartAddr 5622a3050c2SJay } 5632a3050c2SJay 5642a3050c2SJay f3_instr_valid := Mux(f3_lastHalf.valid,f3_hasHalfValid ,VecInit(f3_pd.map(inst => inst.valid))) 5652a3050c2SJay 5662a3050c2SJay /*** frontend Trigger ***/ 5672a3050c2SJay frontendTrigger.io.pds := f3_pd 5682a3050c2SJay frontendTrigger.io.pc := f3_pc 5692a3050c2SJay frontendTrigger.io.data := f3_cut_data 5702a3050c2SJay 5712a3050c2SJay frontendTrigger.io.frontendTrigger := io.frontendTrigger 5722a3050c2SJay frontendTrigger.io.csrTriggerEnable := io.csrTriggerEnable 5732a3050c2SJay 5742a3050c2SJay val f3_triggered = frontendTrigger.io.triggered 5752a3050c2SJay 5762a3050c2SJay /*** send to Ibuffer ***/ 5772a3050c2SJay 5782a3050c2SJay io.toIbuffer.valid := f3_valid && (!f3_req_is_mmio || f3_mmio_can_go) && !f3_flush 5792a3050c2SJay io.toIbuffer.bits.instrs := f3_expd_instr 5802a3050c2SJay io.toIbuffer.bits.valid := f3_instr_valid.asUInt 5815995c9e7SJenius io.toIbuffer.bits.enqEnable := checkerOutStage1.fixedRange.asUInt & f3_instr_valid.asUInt 5822a3050c2SJay io.toIbuffer.bits.pd := f3_pd 58309c6f1ddSLingrui98 io.toIbuffer.bits.ftqPtr := f3_ftq_req.ftqIdx 5842a3050c2SJay io.toIbuffer.bits.pc := f3_pc 5855995c9e7SJenius io.toIbuffer.bits.ftqOffset.zipWithIndex.map{case(a, i) => a.bits := i.U; a.valid := checkerOutStage1.fixedTaken(i) && !f3_req_is_mmio} 5862a3050c2SJay io.toIbuffer.bits.foldpc := f3_foldpc 5873908fff2SJay io.toIbuffer.bits.ipf := VecInit(f3_pf_vec.zip(f3_crossPageFault).map{case (pf, crossPF) => pf || crossPF}) 5882a3050c2SJay io.toIbuffer.bits.acf := f3_af_vec 5892a3050c2SJay io.toIbuffer.bits.crossPageIPFFix := f3_crossPageFault 5902a3050c2SJay io.toIbuffer.bits.triggered := f3_triggered 5912a3050c2SJay 5922a3050c2SJay when(f3_lastHalf.valid){ 5935995c9e7SJenius io.toIbuffer.bits.enqEnable := checkerOutStage1.fixedRange.asUInt & f3_instr_valid.asUInt & f3_lastHalf_mask 5942a3050c2SJay io.toIbuffer.bits.valid := f3_lastHalf_mask & f3_instr_valid.asUInt 5952a3050c2SJay } 5962a3050c2SJay 5972a3050c2SJay /** external predecode for MMIO instruction */ 5982a3050c2SJay when(f3_req_is_mmio){ 599ee175d78SJay val inst = Cat(f3_mmio_data(1), f3_mmio_data(0)) 6002a3050c2SJay val currentIsRVC = isRVC(inst) 6012a3050c2SJay 6022a3050c2SJay val brType::isCall::isRet::Nil = brInfo(inst) 6032a3050c2SJay val jalOffset = jal_offset(inst, currentIsRVC) 6042a3050c2SJay val brOffset = br_offset(inst, currentIsRVC) 6052a3050c2SJay 6062a3050c2SJay io.toIbuffer.bits.instrs (0) := new RVCDecoder(inst, XLEN).decode.bits 6072a3050c2SJay 6082a3050c2SJay io.toIbuffer.bits.pd(0).valid := true.B 6092a3050c2SJay io.toIbuffer.bits.pd(0).isRVC := currentIsRVC 6102a3050c2SJay io.toIbuffer.bits.pd(0).brType := brType 6112a3050c2SJay io.toIbuffer.bits.pd(0).isCall := isCall 6122a3050c2SJay io.toIbuffer.bits.pd(0).isRet := isRet 6132a3050c2SJay 614ee175d78SJay io.toIbuffer.bits.acf(0) := mmio_resend_af 615c3b2d83aSJay io.toIbuffer.bits.ipf(0) := mmio_resend_pf 616c3b2d83aSJay io.toIbuffer.bits.crossPageIPFFix(0) := mmio_resend_pf 617ee175d78SJay 6182a3050c2SJay io.toIbuffer.bits.enqEnable := f3_mmio_range.asUInt 6192a3050c2SJay } 6202a3050c2SJay 62109c6f1ddSLingrui98 62209c6f1ddSLingrui98 //Write back to Ftq 623a37fbf10SJay val f3_cache_fetch = f3_valid && !(f2_fire && !f2_flush) 624a37fbf10SJay val finishFetchMaskReg = RegNext(f3_cache_fetch) 625a37fbf10SJay 6262a3050c2SJay val mmioFlushWb = Wire(Valid(new PredecodeWritebackBundle)) 6270be662e4SJay val f3_mmio_missOffset = Wire(ValidUndirectioned(UInt(log2Ceil(PredictWidth).W))) 628a37fbf10SJay f3_mmio_missOffset.valid := f3_req_is_mmio 6290be662e4SJay f3_mmio_missOffset.bits := 0.U 6300be662e4SJay 631ee175d78SJay mmioFlushWb.valid := (f3_req_is_mmio && mmio_state === m_waitCommit && RegNext(fromUncache.fire()) && f3_mmio_use_seq_pc) 6322a3050c2SJay mmioFlushWb.bits.pc := f3_pc 6332a3050c2SJay mmioFlushWb.bits.pd := f3_pd 6342a3050c2SJay mmioFlushWb.bits.pd.zipWithIndex.map{case(instr,i) => instr.valid := f3_mmio_range(i)} 6352a3050c2SJay mmioFlushWb.bits.ftqIdx := f3_ftq_req.ftqIdx 6362a3050c2SJay mmioFlushWb.bits.ftqOffset := f3_ftq_req.ftqOffset.bits 6372a3050c2SJay mmioFlushWb.bits.misOffset := f3_mmio_missOffset 6382a3050c2SJay mmioFlushWb.bits.cfiOffset := DontCare 639ee175d78SJay mmioFlushWb.bits.target := Mux(mmio_is_RVC, f3_ftq_req.startAddr + 2.U , f3_ftq_req.startAddr + 4.U) 6402a3050c2SJay mmioFlushWb.bits.jalTarget := DontCare 6412a3050c2SJay mmioFlushWb.bits.instrRange := f3_mmio_range 64209c6f1ddSLingrui98 643ee175d78SJay mmio_redirect := (f3_req_is_mmio && mmio_state === m_waitCommit && RegNext(fromUncache.fire()) && f3_mmio_use_seq_pc) 64409c6f1ddSLingrui98 64500240ba6SJay XSPerfAccumulate("fetch_bubble_ibuffer_not_ready", io.toIbuffer.valid && !io.toIbuffer.ready ) 64600240ba6SJay 64700240ba6SJay 64858dbdfc2SJay /** 64958dbdfc2SJay ****************************************************************************** 65058dbdfc2SJay * IFU Write Back Stage 65158dbdfc2SJay * - write back predecode information to Ftq to update 65258dbdfc2SJay * - redirect if found fault prediction 65358dbdfc2SJay * - redirect if has false hit last half (last PC is not start + 32 Bytes, but in the midle of an notCFI RVI instruction) 65458dbdfc2SJay ****************************************************************************** 6552a3050c2SJay */ 65658dbdfc2SJay 6572a3050c2SJay val wb_valid = RegNext(RegNext(f2_fire && !f2_flush) && !f3_req_is_mmio && !f3_flush) 6582a3050c2SJay val wb_ftq_req = RegNext(f3_ftq_req) 659cd365d4cSrvcoresjw 6605995c9e7SJenius val wb_check_result_stage1 = RegNext(checkerOutStage1) 6615995c9e7SJenius val wb_check_result_stage2 = checkerOutStage2 6622a3050c2SJay val wb_instr_range = RegNext(io.toIbuffer.bits.enqEnable) 6632a3050c2SJay val wb_pc = RegNext(f3_pc) 6642a3050c2SJay val wb_pd = RegNext(f3_pd) 6652a3050c2SJay val wb_instr_valid = RegNext(f3_instr_valid) 6662a3050c2SJay 6672a3050c2SJay /* false hit lastHalf */ 6682a3050c2SJay val wb_lastIdx = RegNext(f3_last_validIdx) 6692a3050c2SJay val wb_false_lastHalf = RegNext(f3_false_lastHalf) && wb_lastIdx =/= (PredictWidth - 1).U 6702a3050c2SJay val wb_false_target = RegNext(f3_false_snpc) 6712a3050c2SJay 6722a3050c2SJay val wb_half_flush = wb_false_lastHalf 6732a3050c2SJay val wb_half_target = wb_false_target 6742a3050c2SJay 675a1351e5dSJay /* false oversize */ 676a1351e5dSJay val lastIsRVC = wb_instr_range.asTypeOf(Vec(PredictWidth,Bool())).last && wb_pd.last.isRVC 677a1351e5dSJay val lastIsRVI = wb_instr_range.asTypeOf(Vec(PredictWidth,Bool()))(PredictWidth - 2) && !wb_pd(PredictWidth - 2).isRVC 6785995c9e7SJenius val lastTaken = wb_check_result_stage1.fixedTaken.last 679a1351e5dSJay 6802a3050c2SJay f3_wb_not_flush := wb_ftq_req.ftqIdx === f3_ftq_req.ftqIdx && f3_valid && wb_valid 6812a3050c2SJay 6823f785aa3SJenius /** if a req with a last half but miss predicted enters in wb stage, and this cycle f3 stalls, 6833f785aa3SJenius * we set a flag to notify f3 that the last half flag need not to be set. 6843f785aa3SJenius */ 685804985a5SJenius //f3_fire is after wb_valid 686*076dea5fSJenius when(wb_valid && RegNext(f3_hasLastHalf,init = false.B) 687*076dea5fSJenius && wb_check_result_stage2.fixedMissPred(PredictWidth - 1) && !f3_fire && !RegNext(f3_fire,init = false.B) 6883f785aa3SJenius ){ 6893f785aa3SJenius f3_lastHalf_disable := true.B 690ab6202e2SJenius } 691ab6202e2SJenius 692804985a5SJenius //wb_valid and f3_fire are in same cycle 693*076dea5fSJenius when(wb_valid && RegNext(f3_hasLastHalf,init = false.B) 694*076dea5fSJenius && wb_check_result_stage2.fixedMissPred(PredictWidth - 1) && f3_fire 695804985a5SJenius ){ 696804985a5SJenius f3_lastHalf.valid := false.B 697804985a5SJenius } 698804985a5SJenius 6992a3050c2SJay val checkFlushWb = Wire(Valid(new PredecodeWritebackBundle)) 7002a3050c2SJay checkFlushWb.valid := wb_valid 7012a3050c2SJay checkFlushWb.bits.pc := wb_pc 7022a3050c2SJay checkFlushWb.bits.pd := wb_pd 7032a3050c2SJay checkFlushWb.bits.pd.zipWithIndex.map{case(instr,i) => instr.valid := wb_instr_valid(i)} 7042a3050c2SJay checkFlushWb.bits.ftqIdx := wb_ftq_req.ftqIdx 7052a3050c2SJay checkFlushWb.bits.ftqOffset := wb_ftq_req.ftqOffset.bits 7065995c9e7SJenius checkFlushWb.bits.misOffset.valid := ParallelOR(wb_check_result_stage2.fixedMissPred) || wb_half_flush 7075995c9e7SJenius checkFlushWb.bits.misOffset.bits := Mux(wb_half_flush, wb_lastIdx, ParallelPriorityEncoder(wb_check_result_stage2.fixedMissPred)) 7085995c9e7SJenius checkFlushWb.bits.cfiOffset.valid := ParallelOR(wb_check_result_stage1.fixedTaken) 7095995c9e7SJenius checkFlushWb.bits.cfiOffset.bits := ParallelPriorityEncoder(wb_check_result_stage1.fixedTaken) 7105995c9e7SJenius checkFlushWb.bits.target := Mux(wb_half_flush, wb_half_target, wb_check_result_stage2.fixedTarget(ParallelPriorityEncoder(wb_check_result_stage2.fixedMissPred))) 7115995c9e7SJenius checkFlushWb.bits.jalTarget := wb_check_result_stage2.fixedTarget(ParallelPriorityEncoder(VecInit(wb_pd.zip(wb_instr_valid).map{case (pd, v) => v && pd.isJal }))) 7122a3050c2SJay checkFlushWb.bits.instrRange := wb_instr_range.asTypeOf(Vec(PredictWidth, Bool())) 7132a3050c2SJay 714bccc5520SJenius toFtq.pdWb := Mux(wb_valid, checkFlushWb, mmioFlushWb) 7152a3050c2SJay 7162a3050c2SJay wb_redirect := checkFlushWb.bits.misOffset.valid && wb_valid 71709c6f1ddSLingrui98 7185b3c20f7SJinYue /*write back flush type*/ 7195995c9e7SJenius val checkFaultType = wb_check_result_stage2.faultType 7205b3c20f7SJinYue val checkJalFault = wb_valid && checkFaultType.map(_.isjalFault).reduce(_||_) 7215b3c20f7SJinYue val checkRetFault = wb_valid && checkFaultType.map(_.isRetFault).reduce(_||_) 7225b3c20f7SJinYue val checkTargetFault = wb_valid && checkFaultType.map(_.istargetFault).reduce(_||_) 7235b3c20f7SJinYue val checkNotCFIFault = wb_valid && checkFaultType.map(_.notCFIFault).reduce(_||_) 7245b3c20f7SJinYue val checkInvalidTaken = wb_valid && checkFaultType.map(_.invalidTakenFault).reduce(_||_) 7255b3c20f7SJinYue 7265b3c20f7SJinYue 7275b3c20f7SJinYue XSPerfAccumulate("predecode_flush_jalFault", checkJalFault ) 7285b3c20f7SJinYue XSPerfAccumulate("predecode_flush_retFault", checkRetFault ) 7295b3c20f7SJinYue XSPerfAccumulate("predecode_flush_targetFault", checkTargetFault ) 7305b3c20f7SJinYue XSPerfAccumulate("predecode_flush_notCFIFault", checkNotCFIFault ) 7315b3c20f7SJinYue XSPerfAccumulate("predecode_flush_incalidTakenFault", checkInvalidTaken ) 7325b3c20f7SJinYue 7335b3c20f7SJinYue when(checkRetFault){ 7345b3c20f7SJinYue XSDebug("startAddr:%x nextstartAddr:%x taken:%d takenIdx:%d\n", 7355b3c20f7SJinYue wb_ftq_req.startAddr, wb_ftq_req.nextStartAddr, wb_ftq_req.ftqOffset.valid, wb_ftq_req.ftqOffset.bits) 7365b3c20f7SJinYue } 7375b3c20f7SJinYue 7381d8f4dcbSJay /** performance counter */ 739005e809bSJiuyang Liu val f3_perf_info = RegEnable(f2_perf_info, f2_fire) 7401d8f4dcbSJay val f3_req_0 = io.toIbuffer.fire() 7411d8f4dcbSJay val f3_req_1 = io.toIbuffer.fire() && f3_doubleLine 7421d8f4dcbSJay val f3_hit_0 = io.toIbuffer.fire() && f3_perf_info.bank_hit(0) 7431d8f4dcbSJay val f3_hit_1 = io.toIbuffer.fire() && f3_doubleLine & f3_perf_info.bank_hit(1) 7441d8f4dcbSJay val f3_hit = f3_perf_info.hit 745cd365d4cSrvcoresjw val perfEvents = Seq( 7462a3050c2SJay ("frontendFlush ", wb_redirect ), 747cd365d4cSrvcoresjw ("ifu_req ", io.toIbuffer.fire() ), 7481d8f4dcbSJay ("ifu_miss ", io.toIbuffer.fire() && !f3_perf_info.hit ), 749cd365d4cSrvcoresjw ("ifu_req_cacheline_0 ", f3_req_0 ), 750cd365d4cSrvcoresjw ("ifu_req_cacheline_1 ", f3_req_1 ), 751cd365d4cSrvcoresjw ("ifu_req_cacheline_0_hit ", f3_hit_1 ), 752cd365d4cSrvcoresjw ("ifu_req_cacheline_1_hit ", f3_hit_1 ), 7531d8f4dcbSJay ("only_0_hit ", f3_perf_info.only_0_hit && io.toIbuffer.fire() ), 7541d8f4dcbSJay ("only_0_miss ", f3_perf_info.only_0_miss && io.toIbuffer.fire() ), 7551d8f4dcbSJay ("hit_0_hit_1 ", f3_perf_info.hit_0_hit_1 && io.toIbuffer.fire() ), 7561d8f4dcbSJay ("hit_0_miss_1 ", f3_perf_info.hit_0_miss_1 && io.toIbuffer.fire() ), 7571d8f4dcbSJay ("miss_0_hit_1 ", f3_perf_info.miss_0_hit_1 && io.toIbuffer.fire() ), 7581d8f4dcbSJay ("miss_0_miss_1 ", f3_perf_info.miss_0_miss_1 && io.toIbuffer.fire() ), 759cd365d4cSrvcoresjw ) 7601ca0e4f3SYinan Xu generatePerfEvent() 76109c6f1ddSLingrui98 762f7c29b0aSJinYue XSPerfAccumulate("ifu_req", io.toIbuffer.fire() ) 763f7c29b0aSJinYue XSPerfAccumulate("ifu_miss", io.toIbuffer.fire() && !f3_hit ) 764f7c29b0aSJinYue XSPerfAccumulate("ifu_req_cacheline_0", f3_req_0 ) 765f7c29b0aSJinYue XSPerfAccumulate("ifu_req_cacheline_1", f3_req_1 ) 766f7c29b0aSJinYue XSPerfAccumulate("ifu_req_cacheline_0_hit", f3_hit_0 ) 767f7c29b0aSJinYue XSPerfAccumulate("ifu_req_cacheline_1_hit", f3_hit_1 ) 7682a3050c2SJay XSPerfAccumulate("frontendFlush", wb_redirect ) 7691d8f4dcbSJay XSPerfAccumulate("only_0_hit", f3_perf_info.only_0_hit && io.toIbuffer.fire() ) 7701d8f4dcbSJay XSPerfAccumulate("only_0_miss", f3_perf_info.only_0_miss && io.toIbuffer.fire() ) 7711d8f4dcbSJay XSPerfAccumulate("hit_0_hit_1", f3_perf_info.hit_0_hit_1 && io.toIbuffer.fire() ) 7721d8f4dcbSJay XSPerfAccumulate("hit_0_miss_1", f3_perf_info.hit_0_miss_1 && io.toIbuffer.fire() ) 7731d8f4dcbSJay XSPerfAccumulate("miss_0_hit_1", f3_perf_info.miss_0_hit_1 && io.toIbuffer.fire() ) 7741d8f4dcbSJay XSPerfAccumulate("miss_0_miss_1", f3_perf_info.miss_0_miss_1 && io.toIbuffer.fire() ) 775a108d429SJay XSPerfAccumulate("hit_0_except_1", f3_perf_info.hit_0_except_1 && io.toIbuffer.fire() ) 776a108d429SJay XSPerfAccumulate("miss_0_except_1", f3_perf_info.miss_0_except_1 && io.toIbuffer.fire() ) 777a108d429SJay XSPerfAccumulate("except_0", f3_perf_info.except_0 && io.toIbuffer.fire() ) 77809c6f1ddSLingrui98} 779d558bd61SJenius 780