xref: /XiangShan/src/main/scala/xiangshan/backend/datapath/DataPath.scala (revision 6ccce5705a02a8b2069deb2ee8fb0915f4c0d1dd)
1package xiangshan.backend.datapath
2
3import org.chipsalliance.cde.config.Parameters
4import chisel3._
5import chisel3.util._
6import difftest.{DiffArchFpRegState, DiffArchIntRegState, DiffArchVecRegState, DifftestModule}
7import freechips.rocketchip.diplomacy.{LazyModule, LazyModuleImp}
8import utility._
9import utils.SeqUtils._
10import utils.{XSPerfAccumulate, XSPerfHistogram}
11import xiangshan._
12import xiangshan.backend.BackendParams
13import xiangshan.backend.Bundles._
14import xiangshan.backend.decode.ImmUnion
15import xiangshan.backend.datapath.DataConfig._
16import xiangshan.backend.datapath.RdConfig._
17import xiangshan.backend.issue.{ImmExtractor, IntScheduler, MemScheduler, VfScheduler}
18import xiangshan.backend.issue.EntryBundles._
19import xiangshan.backend.regfile._
20import xiangshan.backend.PcToDataPathIO
21
22class DataPath(params: BackendParams)(implicit p: Parameters) extends LazyModule {
23  override def shouldBeInlined: Boolean = false
24
25  private implicit val dpParams: BackendParams = params
26  lazy val module = new DataPathImp(this)
27
28  println(s"[DataPath] Preg Params: ")
29  println(s"[DataPath]   Int R(${params.getRfReadSize(IntData())}), W(${params.getRfWriteSize(IntData())}) ")
30  println(s"[DataPath]   Vf R(${params.getRfReadSize(VecData())}), W(${params.getRfWriteSize(VecData())}) ")
31}
32
33class DataPathImp(override val wrapper: DataPath)(implicit p: Parameters, params: BackendParams)
34  extends LazyModuleImp(wrapper) with HasXSParameter {
35
36  private val VCONFIG_PORT = params.vconfigPort
37  private val VLD_PORT = params.vldPort
38
39  val io = IO(new DataPathIO())
40
41  private val (fromIntIQ, toIntIQ, toIntExu) = (io.fromIntIQ, io.toIntIQ, io.toIntExu)
42  private val (fromMemIQ, toMemIQ, toMemExu) = (io.fromMemIQ, io.toMemIQ, io.toMemExu)
43  private val (fromVfIQ , toVfIQ , toVfExu ) = (io.fromVfIQ , io.toVfIQ , io.toFpExu)
44
45  println(s"[DataPath] IntIQ(${fromIntIQ.size}), MemIQ(${fromMemIQ.size})")
46  println(s"[DataPath] IntExu(${fromIntIQ.map(_.size).sum}), MemExu(${fromMemIQ.map(_.size).sum})")
47
48  // just refences for convience
49  private val fromIQ: Seq[MixedVec[DecoupledIO[IssueQueueIssueBundle]]] = (fromIntIQ ++ fromVfIQ ++ fromMemIQ).toSeq
50
51  private val toIQs = toIntIQ ++ toVfIQ ++ toMemIQ
52
53  private val toExu: Seq[MixedVec[DecoupledIO[ExuInput]]] = (toIntExu ++ toVfExu ++ toMemExu).toSeq
54
55  private val fromFlattenIQ: Seq[DecoupledIO[IssueQueueIssueBundle]] = fromIQ.flatten
56
57  private val toFlattenExu: Seq[DecoupledIO[ExuInput]] = toExu.flatten
58
59  private val intWbBusyArbiter = Module(new IntRFWBCollideChecker(backendParams))
60  private val vfWbBusyArbiter = Module(new VfRFWBCollideChecker(backendParams))
61  private val intRFReadArbiter = Module(new IntRFReadArbiter(backendParams))
62  private val vfRFReadArbiter = Module(new VfRFReadArbiter(backendParams))
63
64  private val og0FailedVec2: MixedVec[Vec[Bool]] = Wire(MixedVec(fromIQ.map(x => Vec(x.size, Bool())).toSeq))
65  private val og1FailedVec2: MixedVec[Vec[Bool]] = Wire(MixedVec(fromIQ.map(x => Vec(x.size, Bool())).toSeq))
66
67  // port -> win
68  private val intRdArbWinner: Seq2[MixedVec[Bool]] = intRFReadArbiter.io.in.map(_.map(x => MixedVecInit(x.map(_.ready).toSeq)).toSeq).toSeq
69  private val vfRdArbWinner: Seq2[MixedVec[Bool]] = vfRFReadArbiter.io.in.map(_.map(x => MixedVecInit(x.map(_.ready).toSeq)).toSeq).toSeq
70  private val intWbNotBlock: Seq[MixedVec[Bool]] = intWbBusyArbiter.io.in.map(x => MixedVecInit(x.map(_.ready).toSeq)).toSeq
71  private val vfWbNotBlock: Seq[MixedVec[Bool]] = vfWbBusyArbiter.io.in.map(x => MixedVecInit(x.map(_.ready).toSeq)).toSeq
72
73  private val intRdNotBlock: Seq2[Bool] = intRdArbWinner.map(_.map(_.asUInt.andR))
74  private val vfRdNotBlock: Seq2[Bool] = vfRdArbWinner.map(_.map(_.asUInt.andR))
75
76  private val intRFReadReq: Seq3[ValidIO[RfReadPortWithConfig]] = fromIQ.map(x => x.map(xx => xx.bits.getIntRfReadValidBundle(xx.valid)).toSeq).toSeq
77  private val intDataSources: Seq[Seq[Vec[DataSource]]] = fromIQ.map(x => x.map(xx => xx.bits.common.dataSources).toSeq)
78  private val intNumRegSrcs: Seq[Seq[Int]] = fromIQ.map(x => x.map(xx => xx.bits.exuParams.numRegSrc).toSeq)
79
80  intRFReadArbiter.io.in.zip(intRFReadReq).zipWithIndex.foreach { case ((arbInSeq2, inRFReadReqSeq2), iqIdx) =>
81    arbInSeq2.zip(inRFReadReqSeq2).zipWithIndex.foreach { case ((arbInSeq, inRFReadReqSeq), exuIdx) =>
82      val srcIndices: Seq[Int] = fromIQ(iqIdx)(exuIdx).bits.exuParams.getRfReadSrcIdx(IntData())
83      for (srcIdx <- 0 until fromIQ(iqIdx)(exuIdx).bits.exuParams.numRegSrc) {
84        if (srcIndices.contains(srcIdx) && inRFReadReqSeq.isDefinedAt(srcIdx)) {
85          if (intNumRegSrcs(iqIdx)(exuIdx) == 2) {
86            val src0Req = inRFReadReqSeq(0).valid && intDataSources(iqIdx)(exuIdx)(0).readReg
87            val src1Req = inRFReadReqSeq(1).valid && intDataSources(iqIdx)(exuIdx)(1).readReg
88            if (srcIdx == 0) {
89              arbInSeq(srcIdx).valid := src0Req || src1Req
90              arbInSeq(srcIdx).bits.addr := Mux(src1Req && !src0Req, inRFReadReqSeq(1).bits.addr,inRFReadReqSeq(0).bits.addr)
91            } else {
92              arbInSeq(srcIdx).valid := src0Req && src1Req
93              arbInSeq(srcIdx).bits.addr := inRFReadReqSeq(srcIdx).bits.addr
94            }
95          } else {
96            arbInSeq(srcIdx).valid := inRFReadReqSeq(srcIdx).valid && intDataSources(iqIdx)(exuIdx)(srcIdx).readReg
97            arbInSeq(srcIdx).bits.addr := inRFReadReqSeq(srcIdx).bits.addr
98          }
99        } else {
100          arbInSeq(srcIdx).valid := false.B
101          arbInSeq(srcIdx).bits.addr := 0.U
102        }
103      }
104    }
105  }
106
107  private val vfRFReadReq: Seq3[ValidIO[RfReadPortWithConfig]] = fromIQ.map(x => x.map(xx => xx.bits.getVfRfReadValidBundle(xx.valid)).toSeq).toSeq
108
109  vfRFReadArbiter.io.in.zip(vfRFReadReq).zipWithIndex.foreach { case ((arbInSeq2, inRFReadReqSeq2), iqIdx) =>
110    arbInSeq2.zip(inRFReadReqSeq2).zipWithIndex.foreach { case ((arbInSeq, inRFReadReqSeq), exuIdx) =>
111      val srcIndices: Seq[Int] = VfRegSrcDataSet.flatMap(data => fromIQ(iqIdx)(exuIdx).bits.exuParams.getRfReadSrcIdx(data)).toSeq.sorted
112      for (srcIdx <- 0 until fromIQ(iqIdx)(exuIdx).bits.exuParams.numRegSrc) {
113        if (srcIndices.contains(srcIdx) && inRFReadReqSeq.isDefinedAt(srcIdx)) {
114          arbInSeq(srcIdx).valid := inRFReadReqSeq(srcIdx).valid
115          arbInSeq(srcIdx).bits.addr := inRFReadReqSeq(srcIdx).bits.addr
116        } else {
117          arbInSeq(srcIdx).valid := false.B
118          arbInSeq(srcIdx).bits.addr := 0.U
119        }
120      }
121    }
122  }
123
124  private val intRFWriteReq: Seq2[Bool] = fromIQ.map(x => x.map(xx => xx.valid && xx.bits.common.rfWen.getOrElse(false.B)).toSeq).toSeq
125  private val vfRFWriteReq: Seq2[Bool] = fromIQ.map(x => x.map(xx => xx.valid && xx.bits.common.getVfWen.getOrElse(false.B)).toSeq).toSeq
126
127  intWbBusyArbiter.io.in.zip(intRFWriteReq).foreach { case (arbInSeq, inRFWriteReqSeq) =>
128    arbInSeq.zip(inRFWriteReqSeq).foreach { case (arbIn, inRFWriteReq) =>
129      arbIn.valid := inRFWriteReq
130    }
131  }
132
133  vfWbBusyArbiter.io.in.zip(vfRFWriteReq).foreach { case (arbInSeq, inRFWriteReqSeq) =>
134    arbInSeq.zip(inRFWriteReqSeq).foreach { case (arbIn, inRFWriteReq) =>
135      arbIn.valid := inRFWriteReq
136    }
137  }
138
139  private val intSchdParams = params.schdParams(IntScheduler())
140  private val vfSchdParams = params.schdParams(VfScheduler())
141  private val memSchdParams = params.schdParams(MemScheduler())
142
143  private val numIntRfReadByExu = intSchdParams.numIntRfReadByExu + memSchdParams.numIntRfReadByExu
144  private val numVfRfReadByExu = vfSchdParams.numVfRfReadByExu + memSchdParams.numVfRfReadByExu
145  // Todo: limit read port
146  private val numIntR = numIntRfReadByExu
147  private val numVfR = numVfRfReadByExu
148  println(s"[DataPath] RegFile read req needed by Exu: Int(${numIntRfReadByExu}), Vf(${numVfRfReadByExu})")
149  println(s"[DataPath] RegFile read port: Int(${numIntR}), Vf(${numVfR})")
150
151  private val schdParams = params.allSchdParams
152
153  private val pcReadValid = Wire(chiselTypeOf(io.fromPcTargetMem.fromDataPathValid))
154  private val pcReadFtqPtr = Wire(chiselTypeOf(io.fromPcTargetMem.fromDataPathFtqPtr))
155  private val pcReadFtqOffset = Wire(chiselTypeOf(io.fromPcTargetMem.fromDataPathFtqOffset))
156  private val targetPCRdata = io.fromPcTargetMem.toDataPathTargetPC
157  private val pcRdata = io.fromPcTargetMem.toDataPathPC
158  private val intRfRaddr = Wire(Vec(params.numPregRd(IntData()), UInt(intSchdParams.pregIdxWidth.W)))
159  private val intRfRdata = Wire(Vec(params.numPregRd(IntData()), UInt(intSchdParams.rfDataWidth.W)))
160  private val intRfWen = Wire(Vec(io.fromIntWb.length, Bool()))
161  private val intRfWaddr = Wire(Vec(io.fromIntWb.length, UInt(intSchdParams.pregIdxWidth.W)))
162  private val intRfWdata = Wire(Vec(io.fromIntWb.length, UInt(intSchdParams.rfDataWidth.W)))
163
164  private val vfRfSplitNum = VLEN / XLEN
165  private val vfRfRaddr = Wire(Vec(params.numPregRd(VecData()), UInt(vfSchdParams.pregIdxWidth.W)))
166  private val vfRfRdata = Wire(Vec(params.numPregRd(VecData()), UInt(vfSchdParams.rfDataWidth.W)))
167  private val vfRfWen = Wire(Vec(vfRfSplitNum, Vec(io.fromVfWb.length, Bool())))
168  private val vfRfWaddr = Wire(Vec(io.fromVfWb.length, UInt(vfSchdParams.pregIdxWidth.W)))
169  private val vfRfWdata = Wire(Vec(io.fromVfWb.length, UInt(vfSchdParams.rfDataWidth.W)))
170
171  val pcReadFtqPtrFormIQ = fromIntIQ.flatten.filter(x => x.bits.exuParams.needPc)
172  assert(pcReadFtqPtrFormIQ.size == pcReadFtqPtr.size, s"pcReadFtqPtrFormIQ.size ${pcReadFtqPtrFormIQ.size} not equal pcReadFtqPtr.size ${pcReadFtqPtr.size}")
173  pcReadValid.zip(pcReadFtqPtrFormIQ.map(_.valid)).map(x => x._1 := x._2)
174  pcReadFtqPtr.zip(pcReadFtqPtrFormIQ.map(_.bits.common.ftqIdx.get)).map(x => x._1 := x._2)
175  pcReadFtqOffset.zip(pcReadFtqPtrFormIQ.map(_.bits.common.ftqOffset.get)).map(x => x._1 := x._2)
176  io.fromPcTargetMem.fromDataPathValid := pcReadValid
177  io.fromPcTargetMem.fromDataPathFtqPtr := pcReadFtqPtr
178  io.fromPcTargetMem.fromDataPathFtqOffset := pcReadFtqOffset
179  private val intDebugRead: Option[(Vec[UInt], Vec[UInt])] =
180    if (env.AlwaysBasicDiff || env.EnableDifftest) {
181      Some(Wire(Vec(32, UInt(intSchdParams.pregIdxWidth.W))), Wire(Vec(32, UInt(XLEN.W))))
182    } else { None }
183  private val vfDebugRead: Option[(Vec[UInt], Vec[UInt])] =
184    if (env.AlwaysBasicDiff || env.EnableDifftest) {
185      Some(Wire(Vec(32 + 32 + 1, UInt(vfSchdParams.pregIdxWidth.W))), Wire(Vec(32 + 32 + 1, UInt(VLEN.W))))
186    } else { None }
187
188  private val fpDebugReadData: Option[Vec[UInt]] =
189    if (env.AlwaysBasicDiff || env.EnableDifftest) {
190      Some(Wire(Vec(32, UInt(XLEN.W))))
191    } else { None }
192  private val vecDebugReadData: Option[Vec[UInt]] =
193    if (env.AlwaysBasicDiff || env.EnableDifftest) {
194      Some(Wire(Vec(64, UInt(64.W)))) // v0 = Cat(Vec(1), Vec(0))
195    } else { None }
196  private val vconfigDebugReadData: Option[UInt] =
197    if (env.AlwaysBasicDiff || env.EnableDifftest) {
198      Some(Wire(UInt(64.W)))
199    } else { None }
200
201
202  fpDebugReadData.foreach(_ := vfDebugRead
203    .get._2
204    .slice(0, 32)
205    .map(_(63, 0))
206  ) // fp only used [63, 0]
207  vecDebugReadData.foreach(_ := vfDebugRead
208    .get._2
209    .slice(32, 64)
210    .map(x => Seq(x(63, 0), x(127, 64))).flatten
211  )
212  vconfigDebugReadData.foreach(_ := vfDebugRead
213    .get._2(64)(63, 0)
214  )
215
216  io.debugVconfig.foreach(_ := vconfigDebugReadData.get)
217
218  IntRegFile("IntRegFile", intSchdParams.numPregs, intRfRaddr, intRfRdata, intRfWen, intRfWaddr, intRfWdata,
219    bankNum = 1,
220    debugReadAddr = intDebugRead.map(_._1),
221    debugReadData = intDebugRead.map(_._2))
222  VfRegFile("VfRegFile", vfSchdParams.numPregs, vfRfSplitNum, vfRfRaddr, vfRfRdata, vfRfWen, vfRfWaddr, vfRfWdata,
223    debugReadAddr = vfDebugRead.map(_._1),
224    debugReadData = vfDebugRead.map(_._2))
225
226  intRfWaddr := io.fromIntWb.map(_.addr).toSeq
227  intRfWdata := io.fromIntWb.map(_.data).toSeq
228  intRfWen := io.fromIntWb.map(_.wen).toSeq
229
230  for (portIdx <- intRfRaddr.indices) {
231    if (intRFReadArbiter.io.out.isDefinedAt(portIdx))
232      intRfRaddr(portIdx) := intRFReadArbiter.io.out(portIdx).bits.addr
233    else
234      intRfRaddr(portIdx) := 0.U
235  }
236
237  vfRfWaddr := io.fromVfWb.map(_.addr).toSeq
238  vfRfWdata := io.fromVfWb.map(_.data).toSeq
239  vfRfWen.foreach(_.zip(io.fromVfWb.map(_.wen)).foreach { case (wenSink, wenSource) => wenSink := wenSource } )// Todo: support fp multi-write
240
241  for (portIdx <- vfRfRaddr.indices) {
242    if (vfRFReadArbiter.io.out.isDefinedAt(portIdx))
243      vfRfRaddr(portIdx) := vfRFReadArbiter.io.out(portIdx).bits.addr
244    else
245      vfRfRaddr(portIdx) := 0.U
246  }
247
248  vfRfRaddr(VCONFIG_PORT) := io.vconfigReadPort.addr
249  io.vconfigReadPort.data := vfRfRdata(VCONFIG_PORT)
250  // vfRfRaddr(VLD_PORT) := io.vldReadPort.addr
251  io.vldReadPort.data := DontCare
252
253  intDebugRead.foreach { case (addr, _) =>
254    addr := io.debugIntRat.get
255  }
256
257  vfDebugRead.foreach { case (addr, _) =>
258    addr := io.debugFpRat.get ++ io.debugVecRat.get :+ io.debugVconfigRat.get
259  }
260  println(s"[DataPath] " +
261    s"has intDebugRead: ${intDebugRead.nonEmpty}, " +
262    s"has vfDebugRead: ${vfDebugRead.nonEmpty}")
263
264  val s1_addrOHs = Reg(MixedVec(
265    fromIQ.map(x => MixedVec(x.map(_.bits.addrOH.cloneType).toSeq)).toSeq
266  ))
267  val s1_toExuValid: MixedVec[MixedVec[Bool]] = Reg(MixedVec(
268    toExu.map(x => MixedVec(x.map(_.valid.cloneType).toSeq)).toSeq
269  ))
270  val s1_toExuData: MixedVec[MixedVec[ExuInput]] = Reg(MixedVec(toExu.map(x => MixedVec(x.map(_.bits.cloneType).toSeq)).toSeq))
271  val s1_immInfo = Reg(MixedVec(toExu.map(x => MixedVec(x.map(x => new ImmInfo).toSeq)).toSeq))
272  s1_immInfo.zip(fromIQ).map { case (s1Vec, s0Vec) =>
273    s1Vec.zip(s0Vec).map { case (s1, s0) =>
274      s1.imm := Mux(s0.valid, s0.bits.common.imm, s1.imm)
275      s1.immType := Mux(s0.valid, s0.bits.immType, s1.immType)
276    }
277  }
278  io.og1ImmInfo.zip(s1_immInfo.flatten).map{ case(out, reg) =>
279    out := reg
280  }
281  val s1_toExuReady = Wire(MixedVec(toExu.map(x => MixedVec(x.map(_.ready.cloneType).toSeq))))
282  val s1_srcType: MixedVec[MixedVec[Vec[UInt]]] = MixedVecInit(fromIQ.map(x => MixedVecInit(x.map(xx => RegEnable(xx.bits.srcType, xx.fire)).toSeq)))
283
284  val s1_intPregRData: MixedVec[MixedVec[Vec[UInt]]] = Wire(MixedVec(toExu.map(x => MixedVec(x.map(_.bits.src.cloneType).toSeq))))
285  val s1_vfPregRData: MixedVec[MixedVec[Vec[UInt]]] = Wire(MixedVec(toExu.map(x => MixedVec(x.map(_.bits.src.cloneType).toSeq))))
286
287  val rfrPortConfigs = schdParams.map(_.issueBlockParams).flatten.map(_.exuBlockParams.map(_.rfrPortConfigs))
288
289  println(s"[DataPath] s1_intPregRData.flatten.flatten.size: ${s1_intPregRData.flatten.flatten.size}, intRfRdata.size: ${intRfRdata.size}")
290  s1_intPregRData.foreach(_.foreach(_.foreach(_ := 0.U)))
291  s1_intPregRData.zip(rfrPortConfigs).foreach { case (iqRdata, iqCfg) =>
292      iqRdata.zip(iqCfg).foreach { case (iuRdata, iuCfg) =>
293        val realIuCfg = iuCfg.map(x => if(x.size > 1) x.filter(_.isInstanceOf[IntRD]) else x).flatten
294        assert(iuRdata.size == realIuCfg.size, "iuRdata.size != realIuCfg.size")
295        iuRdata.zip(realIuCfg)
296          .filter { case (_, rfrPortConfig) => rfrPortConfig.isInstanceOf[IntRD] }
297          .foreach { case (sink, cfg) => sink := intRfRdata(cfg.port) }
298      }
299  }
300
301  println(s"[DataPath] s1_vfPregRData.flatten.flatten.size: ${s1_vfPregRData.flatten.flatten.size}, vfRfRdata.size: ${vfRfRdata.size}")
302  s1_vfPregRData.foreach(_.foreach(_.foreach(_ := 0.U)))
303  s1_vfPregRData.zip(rfrPortConfigs).foreach{ case(iqRdata, iqCfg) =>
304      iqRdata.zip(iqCfg).foreach{ case(iuRdata, iuCfg) =>
305        val realIuCfg = iuCfg.map(x => if(x.size > 1) x.filter(_.isInstanceOf[VfRD]) else x).flatten
306        assert(iuRdata.size == realIuCfg.size, "iuRdata.size != realIuCfg.size")
307        iuRdata.zip(realIuCfg)
308          .filter { case (_, rfrPortConfig) => rfrPortConfig.isInstanceOf[VfRD] }
309          .foreach { case (sink, cfg) => sink := vfRfRdata(cfg.port) }
310      }
311  }
312
313  val og0_cancel_no_load = og0FailedVec2.flatten.zip(params.allExuParams).filter(!_._2.hasLoadFu).map(_._1)
314  val og0_cancel_delay = RegNext(VecInit(og0_cancel_no_load.toSeq))
315  for (i <- fromIQ.indices) {
316    for (j <- fromIQ(i).indices) {
317      // IQ(s0) --[Ctrl]--> s1Reg ---------- begin
318      // refs
319      val s1_valid = s1_toExuValid(i)(j)
320      val s1_ready = s1_toExuReady(i)(j)
321      val s1_data = s1_toExuData(i)(j)
322      val s1_addrOH = s1_addrOHs(i)(j)
323      val s0 = fromIQ(i)(j) // s0
324
325      val srcNotBlock = Wire(Bool())
326      srcNotBlock := s0.bits.common.dataSources.zip(intRdArbWinner(i)(j) zip vfRdArbWinner(i)(j)).map { case (source, win) =>
327        !source.readReg || win._1 && win._2
328      }.fold(true.B)(_ && _)
329      if (fromIQ(i)(j).bits.exuParams.schdType.isInstanceOf[IntScheduler] && (fromIQ(i)(j).bits.exuParams.numRegSrc == 2)) {
330        val src0VfBlock = s0.bits.common.dataSources(0).readReg && !vfRdArbWinner(i)(j)(0)
331        val src1VfBlock = s0.bits.common.dataSources(1).readReg && !vfRdArbWinner(i)(j)(1)
332        val src1IntBlock = s0.bits.common.dataSources(0).readReg && s0.bits.common.dataSources(1).readReg && !intRdArbWinner(i)(j)(1)
333        val src0IntBlock = (s0.bits.common.dataSources(0).readReg || s0.bits.common.dataSources(1).readReg) && !intRdArbWinner(i)(j)(0)
334        srcNotBlock := !src0VfBlock && !src1VfBlock && !src1IntBlock && !src0IntBlock
335      }
336      val notBlock = srcNotBlock && intWbNotBlock(i)(j) && vfWbNotBlock(i)(j)
337      val s1_flush = s0.bits.common.robIdx.needFlush(Seq(io.flush, RegNextWithEnable(io.flush)))
338      val s1_cancel = og1FailedVec2(i)(j)
339      val s0_cancel = Wire(Bool())
340      if (s0.bits.exuParams.isIQWakeUpSink) {
341        val exuOHNoLoad = s0.bits.common.l1ExuOH.get.map(x => x.asTypeOf(Vec(x.getWidth, Bool())).zip(params.allExuParams).filter(!_._2.hasLoadFu).map(_._1))
342        s0_cancel := exuOHNoLoad.zip(s0.bits.common.dataSources).map{
343          case (exuOH, dataSource) => (VecInit(exuOH).asUInt & og0_cancel_delay.asUInt).orR && dataSource.readForward
344        }.reduce(_ || _) && s0.valid
345      } else s0_cancel := false.B
346      val s0_ldCancel = LoadShouldCancel(s0.bits.common.loadDependency, io.ldCancel)
347      when (s0.fire && !s1_flush && notBlock && !s1_cancel && !s0_ldCancel && !s0_cancel) {
348        s1_valid := s0.valid
349        s1_data.fromIssueBundle(s0.bits) // no src data here
350        if (fromIQ(i)(j).bits.exuParams.schdType.isInstanceOf[IntScheduler] && (fromIQ(i)(j).bits.exuParams.numRegSrc == 2)) {
351          s1_data.dataSources(1).value := Mux(!s0.bits.common.dataSources(0).readReg && s0.bits.common.dataSources(1).readReg, DataSource.anotherReg, s0.bits.common.dataSources(1).value)
352        }
353        s1_addrOH := s0.bits.addrOH
354      }.otherwise {
355        s1_valid := false.B
356      }
357      s0.ready := (s1_ready || !s1_valid) && notBlock && !s1_cancel && !s0_ldCancel && !s0_cancel
358      // IQ(s0) --[Ctrl]--> s1Reg ---------- end
359    }
360  }
361
362  private val fromIQFire = fromIQ.map(_.map(_.fire))
363  private val toExuFire = toExu.map(_.map(_.fire))
364  toIQs.zipWithIndex.foreach {
365    case(toIQ, iqIdx) =>
366      toIQ.zipWithIndex.foreach {
367        case (toIU, iuIdx) =>
368          // IU: issue unit
369          val og0resp = toIU.og0resp
370          og0FailedVec2(iqIdx)(iuIdx) := fromIQ(iqIdx)(iuIdx).valid && (!fromIQFire(iqIdx)(iuIdx))
371          og0resp.valid                 := og0FailedVec2(iqIdx)(iuIdx)
372          og0resp.bits.robIdx           := fromIQ(iqIdx)(iuIdx).bits.common.robIdx
373          og0resp.bits.uopIdx.foreach(_ := fromIQ(iqIdx)(iuIdx).bits.common.vpu.get.vuopIdx)
374          og0resp.bits.resp             := RespType.block
375          og0resp.bits.fuType           := fromIQ(iqIdx)(iuIdx).bits.common.fuType
376
377          val og1resp = toIU.og1resp
378          og1FailedVec2(iqIdx)(iuIdx)   := s1_toExuValid(iqIdx)(iuIdx) && !toExuFire(iqIdx)(iuIdx)
379          og1resp.valid                 := s1_toExuValid(iqIdx)(iuIdx)
380          og1resp.bits.robIdx           := s1_toExuData(iqIdx)(iuIdx).robIdx
381          og1resp.bits.uopIdx.foreach(_ := s1_toExuData(iqIdx)(iuIdx).vpu.get.vuopIdx)
382          // respType:  fuIdle      ->IQ entry clear
383          //            fuUncertain ->IQ entry no action
384          //            fuBusy      ->IQ entry issued set false, then re-issue
385          // Only hyu, lda and sta are fuUncertain at OG1 stage
386          og1resp.bits.resp             := Mux(!og1FailedVec2(iqIdx)(iuIdx),
387            if (toIU.issueQueueParams match { case x => x.isMemAddrIQ && !x.isVecMemIQ }) RespType.uncertain else RespType.success,
388            RespType.block
389          )
390          og1resp.bits.fuType           := s1_toExuData(iqIdx)(iuIdx).fuType
391      }
392  }
393
394  io.og0CancelOH := VecInit(fromFlattenIQ.map(x => x.valid && !x.fire)).asUInt
395  io.og1CancelOH := VecInit(toFlattenExu.map(x => x.valid && !x.fire)).asUInt
396
397  io.cancelToBusyTable.zipWithIndex.foreach { case (cancel, i) =>
398    cancel.valid := fromFlattenIQ(i).valid && !fromFlattenIQ(i).fire
399    cancel.bits.rfWen := fromFlattenIQ(i).bits.common.rfWen.getOrElse(false.B)
400    cancel.bits.fpWen := fromFlattenIQ(i).bits.common.fpWen.getOrElse(false.B)
401    cancel.bits.vecWen := fromFlattenIQ(i).bits.common.vecWen.getOrElse(false.B)
402    cancel.bits.pdest := fromFlattenIQ(i).bits.common.pdest
403  }
404
405  for (i <- toExu.indices) {
406    for (j <- toExu(i).indices) {
407      // s1Reg --[Ctrl]--> exu(s1) ---------- begin
408      // refs
409      val sinkData = toExu(i)(j).bits
410      // assign
411      toExu(i)(j).valid := s1_toExuValid(i)(j)
412      s1_toExuReady(i)(j) := toExu(i)(j).ready
413      sinkData := s1_toExuData(i)(j)
414      // s1Reg --[Ctrl]--> exu(s1) ---------- end
415
416      // s1Reg --[Data]--> exu(s1) ---------- begin
417      // data source1: preg read data
418      for (k <- sinkData.src.indices) {
419        val srcDataTypeSet: Set[DataConfig] = sinkData.params.getSrcDataType(k)
420
421        val readRfMap: Seq[(Bool, UInt)] = (Seq(None) :+
422          (if (s1_intPregRData(i)(j).isDefinedAt(k) && srcDataTypeSet.intersect(IntRegSrcDataSet).nonEmpty)
423            Some(SrcType.isXp(s1_srcType(i)(j)(k)) -> s1_intPregRData(i)(j)(k))
424          else None) :+
425          (if (s1_vfPregRData(i)(j).isDefinedAt(k) && srcDataTypeSet.intersect(VfRegSrcDataSet).nonEmpty)
426            Some(SrcType.isVfp(s1_srcType(i)(j)(k))-> s1_vfPregRData(i)(j)(k))
427          else None)
428        ).filter(_.nonEmpty).map(_.get)
429        if (readRfMap.nonEmpty)
430          sinkData.src(k) := Mux1H(readRfMap)
431      }
432      if (sinkData.params.hasJmpFu) {
433        val index = pcReadFtqPtrFormIQ.map(_.bits.exuParams).indexOf(sinkData.params)
434        sinkData.pc.get := pcRdata(index)
435      }
436      if (sinkData.params.needTarget) {
437        val index = pcReadFtqPtrFormIQ.map(_.bits.exuParams).indexOf(sinkData.params)
438        sinkData.predictInfo.get.target := targetPCRdata(index)
439      }
440    }
441  }
442
443  if (env.AlwaysBasicDiff || env.EnableDifftest) {
444    val delayedCnt = 2
445    val difftestArchIntRegState = DifftestModule(new DiffArchIntRegState, delay = delayedCnt)
446    difftestArchIntRegState.coreid := io.hartId
447    difftestArchIntRegState.value := intDebugRead.get._2
448
449    val difftestArchFpRegState = DifftestModule(new DiffArchFpRegState, delay = delayedCnt)
450    difftestArchFpRegState.coreid := io.hartId
451    difftestArchFpRegState.value := fpDebugReadData.get
452
453    val difftestArchVecRegState = DifftestModule(new DiffArchVecRegState, delay = delayedCnt)
454    difftestArchVecRegState.coreid := io.hartId
455    difftestArchVecRegState.value := vecDebugReadData.get
456  }
457
458  val int_regcache_size = 48
459  val int_regcache_tag = RegInit(VecInit(Seq.fill(int_regcache_size)(0.U(intSchdParams.pregIdxWidth.W))))
460  val int_regcache_enqPtr = RegInit(0.U(log2Up(int_regcache_size).W))
461  int_regcache_enqPtr := int_regcache_enqPtr + PopCount(intRfWen)
462  for (i <- intRfWen.indices) {
463    when (intRfWen(i)) {
464      int_regcache_tag(int_regcache_enqPtr + PopCount(intRfWen.take(i))) := intRfWaddr(i)
465    }
466  }
467
468  val vf_regcache_size = 48
469  val vf_regcache_tag = RegInit(VecInit(Seq.fill(vf_regcache_size)(0.U(vfSchdParams.pregIdxWidth.W))))
470  val vf_regcache_enqPtr = RegInit(0.U(log2Up(vf_regcache_size).W))
471  vf_regcache_enqPtr := vf_regcache_enqPtr + PopCount(vfRfWen.head)
472  for (i <- vfRfWen.indices) {
473    when (vfRfWen.head(i)) {
474      vf_regcache_tag(vf_regcache_enqPtr + PopCount(vfRfWen.head.take(i))) := vfRfWaddr(i)
475    }
476  }
477
478  XSPerfHistogram(s"IntRegFileRead_hist", PopCount(intRFReadArbiter.io.in.flatten.flatten.map(_.valid)), true.B, 0, 20, 1)
479  XSPerfHistogram(s"VfRegFileRead_hist", PopCount(vfRFReadArbiter.io.in.flatten.flatten.map(_.valid)), true.B, 0, 20, 1)
480  XSPerfHistogram(s"IntRegFileWrite_hist", PopCount(intRFWriteReq.flatten), true.B, 0, 20, 1)
481  XSPerfHistogram(s"VfRegFileWrite_hist", PopCount(vfRFWriteReq.flatten), true.B, 0, 20, 1)
482
483  val int_regcache_part32 = (1 until 33).map(i => int_regcache_tag(int_regcache_enqPtr - i.U))
484  val int_regcache_part24 = (1 until 24).map(i => int_regcache_tag(int_regcache_enqPtr - i.U))
485  val int_regcache_part16 = (1 until 17).map(i => int_regcache_tag(int_regcache_enqPtr - i.U))
486  val int_regcache_part8 = (1 until 9).map(i => int_regcache_tag(int_regcache_enqPtr - i.U))
487
488  val int_regcache_48_hit_vec = intRFReadArbiter.io.in.flatten.flatten.map(x => x.valid && int_regcache_tag.map(_ === x.bits.addr).reduce(_ || _))
489  val int_regcache_8_hit_vec = intRFReadArbiter.io.in.flatten.flatten.map(x => x.valid && int_regcache_part8.map(_ === x.bits.addr).reduce(_ || _))
490  val int_regcache_16_hit_vec = intRFReadArbiter.io.in.flatten.flatten.map(x => x.valid && int_regcache_part16.map(_ === x.bits.addr).reduce(_ || _))
491  val int_regcache_24_hit_vec = intRFReadArbiter.io.in.flatten.flatten.map(x => x.valid && int_regcache_part24.map(_ === x.bits.addr).reduce(_ || _))
492  val int_regcache_32_hit_vec = intRFReadArbiter.io.in.flatten.flatten.map(x => x.valid && int_regcache_part32.map(_ === x.bits.addr).reduce(_ || _))
493  XSPerfAccumulate("IntRegCache48Hit", PopCount(int_regcache_48_hit_vec))
494  XSPerfAccumulate("IntRegCache8Hit", PopCount(int_regcache_8_hit_vec))
495  XSPerfAccumulate("IntRegCache16Hit", PopCount(int_regcache_16_hit_vec))
496  XSPerfAccumulate("IntRegCache24Hit", PopCount(int_regcache_24_hit_vec))
497  XSPerfAccumulate("IntRegCache32Hit", PopCount(int_regcache_32_hit_vec))
498  XSPerfHistogram("IntRegCache48Hit_hist", PopCount(int_regcache_48_hit_vec), true.B, 0, 16, 2)
499
500  XSPerfAccumulate(s"IntRFReadBeforeArb", PopCount(intRFReadArbiter.io.in.flatten.flatten.map(_.valid)))
501  XSPerfAccumulate(s"IntRFReadAfterArb", PopCount(intRFReadArbiter.io.out.map(_.valid)))
502  XSPerfAccumulate(s"VfRFReadBeforeArb", PopCount(vfRFReadArbiter.io.in.flatten.flatten.map(_.valid)))
503  XSPerfAccumulate(s"VfRFReadAfterArb", PopCount(vfRFReadArbiter.io.out.map(_.valid)))
504  XSPerfAccumulate(s"IntUopBeforeArb", PopCount(fromIntIQ.flatten.map(_.valid)))
505  XSPerfAccumulate(s"IntUopAfterArb", PopCount(fromIntIQ.flatten.map(_.fire)))
506  XSPerfAccumulate(s"MemUopBeforeArb", PopCount(fromMemIQ.flatten.map(_.valid)))
507  XSPerfAccumulate(s"MemUopAfterArb", PopCount(fromMemIQ.flatten.map(_.fire)))
508  XSPerfAccumulate(s"VfUopBeforeArb", PopCount(fromVfIQ.flatten.map(_.valid)))
509  XSPerfAccumulate(s"VfUopAfterArb", PopCount(fromVfIQ.flatten.map(_.fire)))
510
511  XSPerfHistogram(s"IntRFReadBeforeArb_hist", PopCount(intRFReadArbiter.io.in.flatten.flatten.map(_.valid)), true.B, 0, 16, 2)
512  XSPerfHistogram(s"IntRFReadAfterArb_hist", PopCount(intRFReadArbiter.io.out.map(_.valid)), true.B, 0, 16, 2)
513  XSPerfHistogram(s"VfRFReadBeforeArb_hist", PopCount(vfRFReadArbiter.io.in.flatten.flatten.map(_.valid)), true.B, 0, 16, 2)
514  XSPerfHistogram(s"VfRFReadAfterArb_hist", PopCount(vfRFReadArbiter.io.out.map(_.valid)), true.B, 0, 16, 2)
515  XSPerfHistogram(s"IntUopBeforeArb_hist", PopCount(fromIntIQ.flatten.map(_.valid)), true.B, 0, 8, 2)
516  XSPerfHistogram(s"IntUopAfterArb_hist", PopCount(fromIntIQ.flatten.map(_.fire)), true.B, 0, 8, 2)
517  XSPerfHistogram(s"MemUopBeforeArb_hist", PopCount(fromMemIQ.flatten.map(_.valid)), true.B, 0, 8, 2)
518  XSPerfHistogram(s"MemUopAfterArb_hist", PopCount(fromMemIQ.flatten.map(_.fire)), true.B, 0, 8, 2)
519  XSPerfHistogram(s"VfUopBeforeArb_hist", PopCount(fromVfIQ.flatten.map(_.valid)), true.B, 0, 8, 2)
520  XSPerfHistogram(s"VfUopAfterArb_hist", PopCount(fromVfIQ.flatten.map(_.fire)), true.B, 0, 8, 2)
521}
522
523class DataPathIO()(implicit p: Parameters, params: BackendParams) extends XSBundle {
524  // params
525  private val intSchdParams = params.schdParams(IntScheduler())
526  private val vfSchdParams = params.schdParams(VfScheduler())
527  private val memSchdParams = params.schdParams(MemScheduler())
528  // bundles
529  val hartId = Input(UInt(8.W))
530
531  val flush: ValidIO[Redirect] = Flipped(ValidIO(new Redirect))
532
533  // Todo: check if this can be removed
534  val vconfigReadPort = new RfReadPort(XLEN, PhyRegIdxWidth)
535
536  val vldReadPort = new RfReadPort(VLEN, PhyRegIdxWidth)
537
538  val wbConfictRead = Input(MixedVec(params.allSchdParams.map(x => MixedVec(x.issueBlockParams.map(x => x.genWbConflictBundle())))))
539
540  val fromIntIQ: MixedVec[MixedVec[DecoupledIO[IssueQueueIssueBundle]]] =
541    Flipped(MixedVec(intSchdParams.issueBlockParams.map(_.genIssueDecoupledBundle)))
542
543  val fromMemIQ: MixedVec[MixedVec[DecoupledIO[IssueQueueIssueBundle]]] =
544    Flipped(MixedVec(memSchdParams.issueBlockParams.map(_.genIssueDecoupledBundle)))
545
546  val fromVfIQ = Flipped(MixedVec(vfSchdParams.issueBlockParams.map(_.genIssueDecoupledBundle)))
547
548  val toIntIQ = MixedVec(intSchdParams.issueBlockParams.map(_.genOGRespBundle))
549
550  val toMemIQ = MixedVec(memSchdParams.issueBlockParams.map(_.genOGRespBundle))
551
552  val toVfIQ = MixedVec(vfSchdParams.issueBlockParams.map(_.genOGRespBundle))
553
554  val og0CancelOH = Output(ExuOH(backendParams.numExu))
555
556  val og1CancelOH = Output(ExuOH(backendParams.numExu))
557
558  val ldCancel = Vec(backendParams.LduCnt + backendParams.HyuCnt, Flipped(new LoadCancelIO))
559
560  val cancelToBusyTable = Vec(backendParams.numExu, ValidIO(new CancelSignal))
561
562  val toIntExu: MixedVec[MixedVec[DecoupledIO[ExuInput]]] = intSchdParams.genExuInputBundle
563
564  val toFpExu: MixedVec[MixedVec[DecoupledIO[ExuInput]]] = MixedVec(vfSchdParams.genExuInputBundle)
565
566  val toMemExu: MixedVec[MixedVec[DecoupledIO[ExuInput]]] = memSchdParams.genExuInputBundle
567
568  val og1ImmInfo: Vec[ImmInfo] = Output(Vec(params.allExuParams.size, new ImmInfo))
569
570  val fromIntWb: MixedVec[RfWritePortWithConfig] = MixedVec(params.genIntWriteBackBundle)
571
572  val fromVfWb: MixedVec[RfWritePortWithConfig] = MixedVec(params.genVfWriteBackBundle)
573
574  val fromPcTargetMem = Flipped(new PcToDataPathIO(params))
575
576  val debugIntRat     = if (params.debugEn) Some(Input(Vec(32, UInt(intSchdParams.pregIdxWidth.W)))) else None
577  val debugFpRat      = if (params.debugEn) Some(Input(Vec(32, UInt(vfSchdParams.pregIdxWidth.W)))) else None
578  val debugVecRat     = if (params.debugEn) Some(Input(Vec(32, UInt(vfSchdParams.pregIdxWidth.W)))) else None
579  val debugVconfigRat = if (params.debugEn) Some(Input(UInt(vfSchdParams.pregIdxWidth.W))) else None
580  val debugVconfig    = if (params.debugEn) Some(Output(UInt(XLEN.W))) else None
581}
582