Login| Sign Up| Help| Contact|

Patent Searching and Data


Title:
ARTIFICIAL INTELLIGENCE FOR CHANNEL STATE INFORMATION
Document Type and Number:
WIPO Patent Application WO/2024/028700
Kind Code:
A1
Abstract:
Various aspects of the present disclosure relate to methods, apparatuses, and systems that support artificial intelligence (AI) for channel state information (CSI). For instance, implementations provide an architecture and associated signaling for compressing an input (e.g., CSI at a user equipment (UE)), quantizing the compressed input, transmitting the quantized compressed input, and extracting (e.g., at a network entity such as a gNB) relevant information from the quantized compressed input. In at least some implementations the architecture includes one or more AI/machine learning (ML) models and is composed of multiple components, such as a UE component and a network entity component.

Inventors:
POURAHMADI VAHID (DE)
HINDY AHMED (US)
KOTHAPALLI VENKATA SRINIVAS (CA)
NANGIA VIJAY (US)
BAGHERI HOSSEIN (US)
Application Number:
PCT/IB2023/057586
Publication Date:
February 08, 2024
Filing Date:
July 26, 2023
Export Citation:
Click for automatic bibliography generation   Help
Assignee:
LENOVO SINGAPORE PTE LTD (SG)
International Classes:
H04B7/06
Domestic Patent References:
WO2021253937A12021-12-23
Other References:
VALENTINA RIZZELLO ET AL: "Learning Representations for CSI Adaptive Quantization and Feedback", ARXIV.ORG, CORNELL UNIVERSITY LIBRARY, 201 OLIN LIBRARY CORNELL UNIVERSITY ITHACA, NY 14853, 13 July 2022 (2022-07-13), XP091271352
LIU ZHENYU ET AL: "An Efficient Deep Learning Framework for Low Rate Massive MIMO CSI Reporting", IEEE TRANSACTIONS ON COMMUNICATIONS, IEEE SERVICE CENTER, PISCATAWAY, NJ. USA, vol. 68, no. 8, 8 May 2020 (2020-05-08), pages 4761 - 4772, XP011804447, ISSN: 0090-6778, [retrieved on 20200813], DOI: 10.1109/TCOMM.2020.2993626
YANG QIANQIAN ET AL: "Deep Convolutional Compression For Massive MIMO CSI Feedback", 2019 IEEE 29TH INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING (MLSP), IEEE, 13 October 2019 (2019-10-13), pages 1 - 6, XP033645807, DOI: 10.1109/MLSP.2019.8918798
3GPP TECHNICAL SPECIFICATION (TS) 38.214
Download PDF:
Claims:
Lenovo Docket No. SMM920220072-WO-PCT 80 What is claimed is: 1. A user equipment (UE) for wireless communication, comprising: at least one memory; and at least one processor coupled with the at least one memory and configured to cause the UE to: generate at least one latent representation of input data based on at least one set of neural network models; generate at least one quantized representation of the at least one latent representation based on at least one of scalar quantization or vector quantization associated with the at least one set of neural network models; and transmit the at least one quantized representation. 2. The UE of claim 1, wherein the at least one processor is configured to cause the UE to determine at least one of a quantization codebook corresponding to the vector quantization, a type of the scalar quantization, or a number of quantization levels for the scalar quantization. 3. The UE of claim 2, wherein the at least one processor is configured to cause the UE to receive an indication from a different apparatus, the indication comprising the at least one of the quantization codebook corresponding to the vector quantization, the type of the scalar quantization, or the number of quantization levels for the scalar quantization. 4. The UE of claim 1, wherein the at least one processor is configured to cause the UE to: determine a first latent representation of the at least one latent representation based on a first set of neural network models; determine a second latent representation of the at least one latent representation based on a second set of neural network models; and wherein: Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 81 a first quantized representation at least one quantized representation is based on vector quantization of the first latent representation, and a second quantized representation of the at least one quantized representation is based on scalar quantization of the second latent representation; and transmit the first quantized representation and the second quantized representation. 5. The UE of claim 4, wherein at the first set of neural network models and the second set of neural network models comprise at least one common neural network model. 6. The UE of claim 1, wherein the at least one processor is configured to cause the UE to determine the at least one set of neural network models from a plurality of set of neural network models based on an indication from a second apparatus. 7. The UE of claim 1, wherein the at least one processor is configured to cause the UE to determine model configuration information for the at least one set of neural network models based on an indication from a second apparatus. 8. The UE of claim 7, wherein the model configuration information comprises at least one of a structure of at least one neural network of the at least one set of neural network models, or weights of at least one neural network of the at least one set of neural network models. 9. The UE of claim 1, wherein the at least one processor is configured to cause the UE to: select the at least one set of neural network models from a plurality of sets of neural network models; and transmit an indication of the selected at least one set of neural network models to a second apparatus. 10. The UE of claim 1, wherein the input data is based at least in part on a channel data representation. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 82 11. The UE of claim 10, wherein least one processor is configured to cause the UE to determine the channel data representation based at least in part on at least one reference signal from a second apparatus. 12. The UE of claim 10, wherein the at least one processor is configured to cause the UE to determine the channel data representation based at least in part on at least one of different transmitter and receiver pairs over different frequency bands, different time slots, or time slot transformation in different domains. 13. The UE of claim 1, wherein the at least one set of neural network models comprises: a first neural network model corresponding to a first configuration of a parameter; and a second neural network model corresponding to a second configuration of the parameter; wherein the parameter comprises at least one of a codebook parameter specifying permitted codewords or a rank parameter specifying permitted ranks associated with the at least one quantized representation. 14. The UE of claim 13, wherein the at least one processor is configured to cause the UE to apply the first neural network model after a time duration from the first configuration of the parameter. 15. A base station for wireless communication, comprising: at least one memory; and at least one processor coupled with the at least one memory and configured to cause the base station to: receive at least one set of data; determine at least one latent representation of the at least one set of data; and determine an output using the at least one latent representation. 16. The base station of claim 15, wherein to determine the at least one latent representation, the at least one processor is configured to cause the base station to at least one of: Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 83 determine at least one of a set of network models, a quantization codebook corresponding to one or more of a vector quantization or a vector dequantization, or a type of scalar quantization; or determine a number of quantization levels for a scalar quantization at least one of directly or based on an indication from a second apparatus, wherein model configuration information for the at least one set of neural network models comprises at least one of a structure of at least one neural network of the at least one set of neural network models, or weights of at least one neural network of the at least one set of neural network models. 17. The base station of claim 16, wherein the at least one processor is configured to cause the base station to determine the model configuration information for the at least one set of neural network models based on an indication received from a second apparatus. 18. The base station of claim 15, wherein the at least one set of data comprises a first set of data and a second set of data, and wherein the at least one processor is configured to cause the base station to: determine a first latent representation of the at least one latent representation based on at least a quantization codebook and the first set of data; determine a second latent representation of the at least one latent representation based on the second set of data; and determine the output using the first latent representation and the second latent representation. 19. A method performed by a user equipment (UE), the method comprising: generating at least one latent representation of input data based on at least one set of neural network models; generating at least one quantized representation of the at least one latent representation based on at least one of scalar quantization or vector quantization associated with the at least one set of neural network models; and transmitting the at least one quantized representation. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 84 20. A processor for wireless comprising: at least one controller coupled with at least one memory and configured to cause the processor to: generate at least one latent representation of input data based on at least one set of neural network models; generate at least one quantized representation of the at least one latent representation based on at least one of scalar quantization or vector quantization associated with the at least one set of neural network models; and transmit the at least one quantized representation. Attorney Docket No. SMM920220072-WO-PCT
Description:
Lenovo Docket No. SMM920220072-WO-PCT 1 ARTIFICIAL INTELLIGENCE FOR CHANNEL STATE INFORMATION RELATED APPLICATIONS [0001] This application claims priority to U.S. Patent Application Serial No. 63/394,814 filed August 3, 2022 entitled “ARTIFICIAL INTELLIGENCE FOR CHANNEL STATE INFORMATION,” U.S. Patent Application Serial No. 63/394,822 filed August 3, 2022 entitled “OPERATION OF A TWO-SIDED MODEL,” and U.S. Patent Application Serial No. 63/394,857 filed August 3, 2022 entitled “GENERATING A MEASUREMENT REPORT USING ONE OF MULTIPLE AVAILABLE ARTIFICIAL INTELLIGENCE MODELS,” the disclosures of which are incorporated by reference herein in their entirety. TECHNICAL FIELD [0002] The present disclosure relates to wireless communications, and more specifically to measurement reports using artificial intelligence (AI) models. BACKGROUND [0003] A wireless communications system may include one or multiple network communication devices, such as base stations, which may be otherwise known as an eNodeB (eNB), a next- generation NodeB (gNB), or other suitable terminology. Each network communication devices, such as a base station may support wireless communications for one or multiple user communication devices, which may be otherwise known as user equipment (UE), or other suitable terminology. The wireless communications system may support wireless communications with one or multiple user communication devices by utilizing resources of the wireless communication system (e.g., time resources (e.g., symbols, slots, subframes, frames, or the like) or frequency resources (e.g., subcarriers, carriers). Additionally, the wireless communications system may support wireless communications across various radio access technologies including third generation (3G) radio access technology, fourth generation (4G) radio access technology, fifth generation (5G) radio access technology, among other suitable radio access technologies beyond 5G (e.g., sixth generation (6G)). Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 2 [0004] In some cases, the wireless system may support measurement and reporting operations. For example, a UE may perform channel measurements and transmit a report, to a base station, such as a channel state information (CSI) report indicating a result of the channel measurements (e.g., CSI). In some cases, one or more of the measurement and reporting operations may occur at various times or in response to various events. SUMMARY [0005] The present disclosure relates to methods, apparatuses, and systems that support AI for CSI. For instance, implementations provide an architecture and associated signaling for compressing an input (e.g., CSI at a UE), quantizing the compressed input, transmitting the quantized compressed input, and extracting (e.g., at a network entity such as a gNB) relevant information from the quantized compressed input. In at least some implementations the architecture includes one or more AI/machine learning (ML) models and is composed of multiple components, such as a UE component and a network entity component. The UE component, for instance, generates two latent representations of input data (e.g., CSI) which are quantized, such as using a scalar quantizer and a codebook quantizer. Both quantized representations are then transferred to the network entity component and the network entity component processes the quantized representations to construct model output, e.g., a reconstruction of the input data processed at the UE component. [0006] By utilizing the described techniques, an amount of data (e.g., CSI-related data) to be exchanged between devices (e.g., UEs and network entities) can be reduced, which can conserve device and network resources. [0007] Some implementations of the methods and apparatuses described herein may further include generating at a first apparatus at least one latent representation of input data based on at least one set of neural network models; generating at least one quantized representation of the at least one latent representation based on at least one of scalar quantization or vector quantization associated with the at least one set of neural network models; and transmitting the at least one quantized representation. [0008] Some implementations of the methods and apparatuses described herein may further include: determining at least one of a quantization codebook corresponding to the vector Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 3 quantization, a type of the scalar quantization, a number of quantization levels for the scalar quantization; receiving an indication from a second apparatus, the indication including the at least one of the quantization codebook corresponding to the vector quantization, the type of the scalar quantization, or the number of quantization levels for the scalar quantization; determining a first latent representation of the at least one latent representation based on a first set of neural network models; determining a second latent representation of the at least one latent representation based on a second set of neural network models; and where: a first quantized representation of the at least one quantized representation is based on vector quantization of the first latent representation, and a second quantized representation of the at least one quantized representation is based on scalar quantization of the second latent representation; and transmitting the first quantized representation and the second quantized representation. [0009] Some implementations of the methods and apparatuses described herein may further include: where at the first set of neural network models and the second set of neural network models include at least one common neural network model; determining the at least one set of neural network models from a plurality of set of neural network models based on an indication from a second apparatus; determining model configuration information for the at least one set of neural network models based on an indication from a second apparatus; where the model configuration information includes at least one of a structure of at least one neural network of the at least one set of neural network models, or weights of at least one neural network of the at least one set of neural network models; selecting the at least one set of neural network models from a plurality of sets of neural network models; and transmitting an indication of the selected at least one set of neural network models to a second apparatus. [0010] Some implementations of the methods and apparatuses described herein may further include: where the input data is based at least in part on a channel data representation; determining the channel data representation based at least in part on at least one reference signal from a second apparatus; determining the channel data representation based at least in part on at least one of different transmitter and receiver pairs over different frequency bands, different time slots, or time slot transformation in different domains; where the at least one set of neural network models includes: a first neural network model corresponding to a first configuration of a parameter; and a second neural network model corresponding to a second configuration of the parameter; where the Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 4 parameter includes at least one of a codebook specifying permitted codewords or a rank parameter specifying permitted ranks associated with the at least one quantized representation; applying the first neural network model after a time duration from the first configuration of the parameter. [0011] Some implementations of the methods and apparatuses described herein may further include receiving at a first apparatus at least one set of data; determining at least one latent representation of the at least one set of data; and determining an output using the at least one latent representation. [0012] Some implementations of the methods and apparatuses described herein may further include: where determining the at least one latent representation includes determining at least one set of neural network models from a plurality of sets of neural network models and based on an indication from a second apparatus; determining model configuration information of the at least one set of neural network models based on an indication received from a second apparatus; where the model configuration information includes at least one of a structure of at least one neural network of the at least one set of neural network models, or weights of at least one neural network of the at least one set of neural network models; where the at least one set of data includes a first set of data and a second set of data, and: determining a first latent representation of the at least one latent representation based on at least a quantization codebook and the first set of data; determining a second latent representation of the at least one latent representation based on the second set of data; and determining the output using the first latent representation and the second latent representation; determining the quantization codebook as a quantization codebook corresponding to a demapper module; determining the quantization codebook based on an indication received from a second apparatus. BRIEF DESCRIPTION OF THE DRAWINGS [0013] FIG. 1 illustrates an example of a wireless communications system that supports AI for CSI in accordance with aspects of the present disclosure. [0014] FIG. 2 illustrates a system that implements at least some CSI feedback mechanisms. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 5 [0015] FIG. 3 illustrates an aperiodic state defining a list of CSI report settings. [0016] FIG. 4 illustrates an information element pertaining to CSI reporting. [0017] FIG. 5 illustrates an information element for RRC configuration for wireless resources. [0018] FIG. 6 illustrates a scenario 600 for partial CSI omission for physical uplink shared channel (PUSCH)-based CSI. [0019] FIGs. 7a and 7b illustrate respectively a UE subsystem and a network subsystem of a CSI system that supports AI for CSI in accordance with aspects of the present disclosure. [0020] FIGs. 8 and 9 illustrate examples of block diagrams of devices that support AI for CSI in accordance with aspects of the present disclosure. [0021] FIGs. 10 through 21 illustrate flowcharts of methods that support AI for CSI in accordance with aspects of the present disclosure. DETAILED DESCRIPTION [0022] In wireless communications systems, A UE can feed back information to a network entity (e.g., a gNB) using a measurement report, such as feeding back CSI information using a CSI measurement report. Given the nature of the information (e.g., real numbers), the measurement reports can be very large. One way to reduce the size of the measurement reports is to compress the data in the measurement reports, such as by using linear compression. However, this compression results typically results in reducing the data size but at the expense of decreased accuracy (or increased distortion) in the data. [0023] Accordingly, this disclosure provides for techniques that support AI for CSI. For instance, implementations provide an architecture and associated signaling for compressing an input (e.g., 3-dimensional (3D) input, CSI at a UE, etc.), quantizing the compressed input, transmitting the quantized compressed input, and extracting (e.g., at a network entity such as a gNB) relevant information from the quantized compressed input. In at least some implementations the architecture includes one or more AI/ML models and is composed of multiple components, such as a UE component and a network entity component. The UE component, for instance, generates two latent representations of input data (e.g., CSI) which are quantized, such as using a scalar quantizer and a Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 6 codebook quantizer. In at least some codewords of the codebook quantizer are not fixed and can be learned during the design of the system. [0024] Both quantized representations are then transferred to the network entity component and the network entity components processes the quantized representations to construct model output, e.g., a reconstruction of the input data processed at the UE component. In some alternative or additional implementations, the described architecture and associated operations can be implemented to use only scalar quantization or only vector quantization (e.g., codebook-based quantization). [0025] Thus, by utilizing the described techniques, an amount of data (e.g., CSI-related data) to be exchanged between devices (e.g., UEs and network entities) can be reduced, which can conserve device and network resources. [0026] Aspects of the present disclosure are described in the context of a wireless communications system. Aspects of the present disclosure are further illustrated and described with reference to device diagrams and flowcharts. [0027] FIG. 1 illustrates an example of a wireless communications system 100 that supports AI for CSI in accordance with aspects of the present disclosure. The wireless communications system 100 may include one or more network entities 102, one or more UE 104, a core network 106, and a packet data network 108. The wireless communications system 100 may support various radio access technologies. In some implementations, the wireless communications system 100 may be a 4G network, such as an LTE network or an LTE-Advanced (LTE-A) network. In some other implementations, the wireless communications system 100 may be a 5G network, such as an NR network. In other implementations, the wireless communications system 100 may be a combination of a 4G network and a 5G network, or other suitable radio access technology including Institute of Electrical and Electronics Engineers (IEEE) 802.11 (Wi-Fi), IEEE 802.16 (WiMAX), IEEE 802.20. The wireless communications system 100 may support radio access technologies beyond 5G. Additionally, the wireless communications system 100 may support technologies, such as time division multiple access (TDMA), frequency division multiple access (FDMA), or code division multiple access (CDMA), etc. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 7 [0028] The one or more network entities may be dispersed throughout a geographic region to form the wireless communications system 100. One or more of the network entities 102 described herein may be or include or may be referred to as a network node, a base station, a network element, a radio access network (RAN), a base transceiver station, an access point, a NodeB, an eNodeB (eNB), a next-generation NodeB (gNB), or other suitable terminology. A network entity 102 and a UE 104 may communicate via a communication link 110, which may be a wireless or wired connection. For example, a network entity 102 and a UE 104 may perform wireless communication (e.g., receive signaling, transmit signaling) over a Uu interface. [0029] A network entity 102 may provide a geographic coverage area 112 for which the network entity 102 may support services (e.g., voice, video, packet data, messaging, broadcast, etc.) for one or more UE 104 within the geographic coverage area 112. For example, a network entity 102 and a UE 104 may support wireless communication of signals related to services (e.g., voice, video, packet data, messaging, broadcast, etc.) according to one or multiple radio access technologies. In some implementations, a network entity 102 may be moveable, for example, a satellite associated with a non-terrestrial network. In some implementations, different geographic coverage areas 112 associated with the same or different radio access technologies may overlap, but the different geographic coverage areas 112 may be associated with different network entities 102. Information and signals described herein may be represented using any of a variety of different technologies and techniques. For example, data, instructions, commands, information, signals, bits, symbols, and chips that may be referenced throughout the description may be represented by voltages, currents, electromagnetic waves, magnetic fields or particles, optical fields or particles, or any combination thereof. [0030] The one or more UE 104 may be dispersed throughout a geographic region of the wireless communications system 100. A UE 104 may include or may be referred to as a mobile device, a wireless device, a remote device, a remote unit, a handheld device, or a subscriber device, or some other suitable terminology. In some implementations, the UE 104 may be referred to as a unit, a station, a terminal, or a client, among other examples. Additionally, or alternatively, the UE 104 may be referred to as an Internet-of-Things (IoT) device, an Internet-of-Everything (IoE) device, or machine-type communication (MTC) device, among other examples. In some Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 8 implementations, a UE 104 may be stationary the wireless communications system 100. In some other implementations, a UE 104 may be mobile in the wireless communications system 100. [0031] The one or more UE 104 may be devices in different forms or having different capabilities. Some examples of UE 104 are illustrated in FIG. 1. A UE 104 may be capable of communicating with various types of devices, such as the network entities 102, other UE 104, or network equipment (e.g., the core network 106, the packet data network 108, a relay device, an integrated access and backhaul (IAB) node, or another network equipment), as shown in FIG. 1. Additionally, or alternatively, a UE 104 may support communication with other network entities 102 or UE 104, which may act as relays in the wireless communications system 100. [0032] A UE 104 may also be able to support wireless communication directly with other UE 104 over a communication link 114. For example, a UE 104 may support wireless communication directly with another UE 104 over a device-to-device (D2D) communication link. In some implementations, such as vehicle-to-vehicle (V2V) deployments, V2X deployments, or cellular- V2X deployments, the communication link 114 may be referred to as a sidelink. For example, a UE 104 may support wireless communication directly with another UE 104 over a PC5 interface. [0033] A network entity 102 may support communications with the core network 106, or with another network entity 102, or both. For example, a network entity 102 may interface with the core network 106 through one or more backhaul links 116 (e.g., via an S1, N2, N2, or another network interface). The network entities 102 may communicate with each other over the backhaul links 116 (e.g., via an X2, Xn, or another network interface). In some implementations, the network entities 102 may communicate with each other directly (e.g., between the network entities 102). In some other implementations, the network entities 102 may communicate with each other or indirectly (e.g., via the core network 106). In some implementations, one or more network entities 102 may include subcomponents, such as an access network entity, which may be an example of an access node controller (ANC). An ANC may communicate with the one or more UE 104 through one or more other access network transmission entities, which may be referred to as a radio heads, smart radio heads, or transmission-reception points (TRPs). [0034] In some implementations, a network entity 102 may be configured in a disaggregated architecture, which may be configured to utilize a protocol stack physically or logically distributed Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 9 among two or more network entities 102, such an integrated access backhaul (IAB) network, an open RAN (O-RAN) (e.g., a network configuration sponsored by the O-RAN Alliance), or a virtualized RAN (vRAN) (e.g., a cloud RAN (C-RAN)). For example, a network entity 102 may include one or more of a central unit (CU), a distributed unit (DU), a radio unit (RU), a RAN Intelligent Controller (RIC) (e.g., a Near-Real Time RIC (Near-real time (RT) RIC), a Non-Real Time RIC (Non-RT RIC)), a Service Management and Orchestration (SMO) system, or any combination thereof. [0035] An RU may also be referred to as a radio head, a smart radio head, a remote radio head (RRH), a remote radio unit (RRU), or a transmission reception point (TRP). One or more components of the network entities 102 in a disaggregated RAN architecture may be co-located, or one or more components of the network entities 102 may be located in distributed locations (e.g., separate physical locations). In some implementations, one or more network entities 102 of a disaggregated RAN architecture may be implemented as virtual units (e.g., a virtual CU (VCU), a virtual DU (VDU), a virtual RU (VRU)). [0036] Split of functionality between a CU, a DU, and an RU may be flexible and may support different functionalities depending upon which functions (e.g., network layer functions, protocol layer functions, baseband functions, radio frequency functions, and any combinations thereof) are performed at a CU, a DU, or an RU. For example, a functional split of a protocol stack may be employed between a CU and a DU such that the CU may support one or more layers of the protocol stack and the DU may support one or more different layers of the protocol stack. In some implementations, the CU may host upper protocol layer (e.g., a layer 3 (L3), a layer 2 (L2)) functionality and signaling (e.g., radio resource control (RRC), service data adaption protocol (SDAP), Packet Data Convergence Protocol (PDCP)). The CU may be connected to one or more DUs or RUs, and the one or more DUs or RUs may host lower protocol layers, such as a layer 1 (L1) (e.g., physical (PHY) layer) or an L2 (e.g., radio link control (RLC) layer, media access control (MAC) layer) functionality and signaling, and may each be at least partially controlled by the CU. [0037] Additionally, or alternatively, a functional split of the protocol stack may be employed between a DU and an RU such that the DU may support one or more layers of the protocol stack and the RU may support one or more different layers of the protocol stack. The DU may support one or multiple different cells (e.g., via one or more RUs). In some implementations, a functional Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 10 split between a CU and a DU, or between a an RU may be within a protocol layer (e.g., some functions for a protocol layer may be performed by one of a CU, a DU, or an RU, while other functions of the protocol layer are performed by a different one of the CU, the DU, or the RU). [0038] A CU may be functionally split further into CU control plane (CU-CP) and CU user plane (CU-UP) functions. A CU may be connected to one or more DUs via a midhaul communication link (e.g., F1, F1-c, F1-u), and a DU may be connected to one or more RUs via a fronthaul communication link (e.g., open fronthaul (FH) interface). In some implementations, a midhaul communication link or a fronthaul communication link may be implemented in accordance with an interface (e.g., a channel) between layers of a protocol stack supported by respective network entities 102 that are in communication via such communication links. [0039] The core network 106 may support user authentication, access authorization, tracking, connectivity, and other access, routing, or mobility functions. The core network 106 may be an evolved packet core (EPC), or a 5G core (5GC), which may include a control plane entity that manages access and mobility (e.g., a mobility management entity (MME), an access and mobility management functions (AMF)) and a user plane entity that routes packets or interconnects to external networks (e.g., a serving gateway (S-GW), a Packet Data Network (PDN) gateway (P- GW), or a user plane function (UPF)). In some implementations, the control plane entity may manage non-access stratum (NAS) functions, such as mobility, authentication, and bearer management (e.g., data bearers, signal bearers, etc.) for the one or more UE 104 served by the one or more network entities 102 associated with the core network 106. [0040] The core network 106 may communicate with the packet data network 108 over one or more backhaul links 116 (e.g., via an S1, N2, N2, or another network interface). The packet data network 108 may include an application server 118. In some implementations, one or more UE 104 may communicate with the application server 118. A UE 104 may establish a session (e.g., a protocol data unit (PDU) session, or the like) with the core network 106 via a network entity 102. The core network 106 may route traffic (e.g., control information, data, and the like) between the UE 104 and the application server 118 using the established session (e.g., the established PDU session). The PDU session may be an example of a logical connection between the UE 104 and the core network 106 (e.g., one or more network functions of the core network 106). Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 11 [0041] In the wireless communications 100, the network entities 102 and the UE 104 may use resources of the wireless communication system 100 (e.g., time resources (e.g., symbols, slots, subframes, frames, or the like) or frequency resources (e.g., subcarriers, carriers) to perform various operations (e.g., wireless communications). In some implementations, the network entities 102 and the UE 104 may support different resource structures. For example, the network entities 102 and the UE 104 may support different frame structures. In some implementations, such as in 4G, the network entities 102 and the UE 104 may support a single frame structure. In some other implementations, such as in 5G and among other suitable radio access technologies, the network entities 102 and the UE 104 may support various frame structures (e.g., multiple frame structures). The network entities 102 and the UE 104 may support various frame structures based on one or more numerologies. [0042] One or more numerologies may be supported in the wireless communications system 100, and a numerology may include a subcarrier spacing and a cyclic prefix. A first numerology (e.g., ^=0) may be associated with a first subcarrier spacing (e.g., 15 kHz) and a normal cyclic prefix. The first numerology (e.g., ^=0) associated with the first subcarrier spacing (e.g., 15 kHz) may utilize one slot per subframe. A second numerology (e.g., ^=1) may be associated with a second subcarrier spacing (e.g., 30 kHz) and a normal cyclic prefix. A third numerology (e.g., ^=2) may be associated with a third subcarrier spacing (e.g., 60 kHz) and a normal cyclic prefix or an extended cyclic prefix. A fourth numerology (e.g., ^=3) may be associated with a fourth subcarrier spacing (e.g., 120 kHz) and a normal cyclic prefix. A fifth numerology (e.g., ^=4) may be associated with a fifth subcarrier spacing (e.g., 240 kHz) and a normal cyclic prefix. [0043] A time interval of a resource (e.g., a communication resource) may be organized according to frames (also referred to as radio frames). Each frame may have a duration, for example, a 10 millisecond (ms) duration. In some implementations, each frame may include multiple subframes. For example, each frame may include 10 subframes, and each subframe may have a duration, for example, a 1 ms duration. In some implementations, each frame may have the same duration. In some implementations, each subframe of a frame may have the same duration. [0044] Additionally or alternatively, a time interval of a resource (e.g., a communication resource) may be organized according to slots. For example, a subframe may include a number (e.g., quantity) of slots. Each slot may include a number (e.g., quantity) of symbols (e.g., orthogonal Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 12 frequency-division multiplexing (OFDM) . In some implementations, the number (e.g., quantity) of slots for a subframe may depend on a numerology. For a normal cyclic prefix, a slot may include 14 symbols. For an extended cyclic prefix (e.g., applicable for 60 kHz subcarrier spacing), a slot may include 12 symbols. The relationship between the number of symbols per slot, the number of slots per subframe, and the number of slots per frame for a normal cyclic prefix and an extended cyclic prefix may depend on a numerology. It should be understood that reference to a first numerology (e.g., ^=0) associated with a first subcarrier spacing (e.g., 15 kHz) may be used interchangeably between subframes and slots. [0045] In the wireless communications system 100, an electromagnetic (EM) spectrum may be split, based on frequency or wavelength, into various classes, frequency bands, frequency channels, etc. By way of example, the wireless communications system 100 may support one or multiple operating frequency bands, such as frequency range designations FR1 (410 MHz – 7.125 GHz), FR2 (24.25 GHz – 52.6 GHz), FR3 (7.125 GHz – 24.25 GHz), FR4 (52.6 GHz – 114.25 GHz), FR4a or FR4-1 (52.6 GHz – 71 GHz), and FR5 (114.25 GHz – 300 GHz). In some implementations, the network entities 102 and the UE 104 may perform wireless communications over one or more of the operating frequency bands. In some implementations, FR1 may be used by the network entities 102 and the UE 104, among other equipment or devices for cellular communications traffic (e.g., control information, data). In some implementations, FR2 may be used by the network entities 102 and the UE 104, among other equipment or devices for short- range, high data rate capabilities. [0046] FR1 may be associated with one or multiple numerologies (e.g., at least three numerologies). For example, FR1 may be associated with a first numerology (e.g., ^=0), which includes 15 kHz subcarrier spacing; a second numerology (e.g., ^=1), which includes 30 kHz subcarrier spacing; and a third numerology (e.g., ^=2), which includes 60 kHz subcarrier spacing. FR2 may be associated with one or multiple numerologies (e.g., at least 2 numerologies). For example, FR2 may be associated with a third numerology (e.g., ^=2), which includes 60 kHz subcarrier spacing; and a fourth numerology (e.g., ^=3), which includes 120 kHz subcarrier spacing. [0047] According to implementations for AI for CSI, a UE 104 and a network entity 102 can interact to enable efficient generation of CSI by the UE 104 and provision of the CSI to the network Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 13 entity 102. For instance, the network entity 102 the UE 104 can exchange AI-CSI information 120 that enables different aspects of AI for CSI to be implemented by the network entity 102 and the UE 104. Examples of the AI-CSI information 120 include configuration information for the UE 104 to measure and report at least one quantity, such as any quantity included in a CSI report. The AI-CSI information 120 can also include different configurations and settings applied by the UE 104, such as part of processing CSI as further described below. [0048] Further, the UE 104 determines CSI for a channel between the network entity 102 and the UE 104 (e.g., based on a reference signal transmitted by the network entity 102) and performs AI-CSI processing 122 to generate processed CSI 126. The AI-CSI processing 122, for instance, includes applying AI processing based on one of multiple AI models 124 (e.g., multiple neural networks) to CSI to generate the processed CSI 126 (e.g., a CSI measurement report). Various details and aspects of the AI-CSI processing 122 and the processed CSI 126 are detailed throughout this disclosure. [0049] The network entity 102 receives the processed CSI 126 and performs AI-CSI processing 128 using the processed CSI 126. For instance, the network entity 102 uses one of multiple AI models 130 (e.g., multiple neural networks) to perform the AI-CSI processing 128 to extract relevant CSI-related data from the processed CSI 126. In at least some implementations the network entity 102 implements portions of the AI-CSI processing 128 utilizing configuration information received from the UE 104, such as part of the AI-CSI information 120. [0050] FIG. 2 illustrates a system 200 that implements at least some CSI feedback mechanisms. The system 200 includes a network entity 102 (e.g., a gNB) equipped with ^ antennas. Further, the system 200 includes UE 104 including a UE 104 1 , a UE 104 2 , and a UE 104 K . In the discussion below, the UE 104 include ^ UE denoted by ^ ^ , ^ ^ , ⋯ , ^ ^ each having ^ antennas. ^ ^ ^ ^^^ denotes a channel at time ^ over frequency band (or subcarrier or subband or physical resource block (PRB) or sub-PRB or PRB-group or bandwidth part in a channel bandwidth) ^, ^ ∈ ^ 1,2, … , ^ ^ , between ^ ^ and ^ ^ which is a matrix of size ^ × ^ with complex entries, e.g., ^ ^ ^ ^^^ ∈ ℂ ^×^ . [0051] At time ^ and frequency band ^, the network entity 102 can be configured to transmit a message ^ ^ ^ ^^^ to user ^ ^ where = ^1,2, ⋯ , ^} while the network entity 102 uses " ^ ^ ^^^ ∈ ℂ ^×^ as the precoding vector. The received signal at ^ ^ , # ^ ^^ ^ ^ , can be written as: Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 14 # ^ ^ ^^^ = + % ^ ^^^^ where % ^ ^^^^ represents the noise vector at the receiver. [0052] Further to the system 200, to improve the achievable rate of the link, the network entity 102 can select " ^ ^ ^^^ that maximizes some metric such as the received signal to noise ratio (SNR). Several schemes have been proposed for optimal selection of " ^ ^ ^^^ where most of them utilize some knowledge about ^ ^ ^ ^^^. [0053] The network entity 102 can obtain information about ^ ^ ^ ^^^ by direct measurement (e.g., in time-division duplexing (TDD) mode and assuming reciprocity of the channel direct measurement of the uplink channel, in frequency-division duplexing (FDD) mode assuming reciprocity of some of the large scale parameters such as AoA/AoD), or indirectly using the information that the UE sends to the gNB (e.g., in FDD mode). In the latter case, a large amount of feedback may be needed to send accurate information about ^ ^ ^ ^^^. [0054] Several methods have been proposed trying to reduce the required CSI feedback. One idea is to try to use the correlation between different entries of ^ ^ ^ ^^), e.g., between a) different Tx- Rx antenna pairs, b) between different frequency bands, c) between different time slots. [0055] In at least some aspects of the present disclosure, implementations consider a single time slot and focus on transmitting information regarding a channel between a user and a network entity over multiple frequency bands. Further, implementations can utilize multiple time slots, such as by replacing a frequency domain with a time domain and/or creating a joint time-frequency domain. For purposes of the discussion herein ^ ^ ^^ ^ ^ may be denoted using ^ ^ . [0056] Further, ^ may be defined as a matrix of size ^ × ^ × ^ which can be constructed by stacking ^ ^ for multiple frequency bands, e.g., the entries at ^&', (, ^) are equal to ^ ^ &', (). Thus, a UE may send information about ^ × ^ × ^ complex numbers to a network entity. [0057] Accordingly, in aspects of this disclosure, implementations enable a UE to transmit information about H to a network entity with a limited number of feedback bits. For instance, implementations to enable a UE to efficiently compress and send CSI information, such as H, to the gNB. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 15 [0058] Thus, implementations discussed enable reduction of feedback information generated by a UE and/or transmitted to a network entity. For instance, data driven schemes for data compression are provided such as by an encoder at a UE which computes a latent representation of input data. The latent representation can be quantized and sent to a network entity and the network entity applies a decoder to the quantized latent representation to reconstruct a desired output. The implementation of different model features such as at a UE and a network entity, and associated architectures, can be referred to herein as two-sided models. Further, implementation details are provided such as for training, deploying, and operating two-sided models in settings that involve different types of UE and/or different network parameters, such as when a number of possible feedback bits may change. [0059] In some wireless communications systems it is considered that a gNB is equipped with a two-dimensional (2D) antenna array with N 1 , N 2 antenna ports per polarization placed horizontally and vertically, and communication occurs over N3 precoding matrix indicator (PMI) sub-bands. A PMI subband consists of a set of resource blocks, each resource block consisting of a set of subcarriers. In such case, 2N 1 N 2 CSI-reference signal (RS) ports are utilized to enable downlink channel estimation with high resolution for NR Rel. 15 Type-II codebook. In order to reduce the uplink feedback overhead, a Discrete Fourier transform (DFT)-based CSI compression of the spatial domain is applied to L dimensions per polarization, where L<N 1 N 2 . In the sequel the indices of the 2L dimensions are referred as the spatial domain (SD) basis indices. The magnitude and phase values of the linear combination coefficients for each sub-band are fed back to the gNB as part of the CSI report. The 2N 1 N 2 xN 3 codebook per layer l takes on the form: * ^ = * ^ * ^,^ , where W 1 is a 2N 1 N 2 x2L block-diagonal with two identical diagonal blocks, e.g.: * + = ./, and B is an N1N2xL matrix with a 2D oversampled DFT matrix, as follows. 567 567 ^ 9 :; 4 4 5 ^ 0 1 = ⋯ <, ? , Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 16 where the Note that O1, O2 oversampling factors are considered for the 2D DFT matrix from which matrix B is drawn. Note that W 1 is common across all layers. W2,l is a 2Lx N3 matrix, where the i th column corresponds to the linear combination coefficients in the i th sub-band. Only the indices of the L selected columns of B may be reported, along with the oversampling index taking on O 1 O 2 values. Note that W2,l are independent for different layers. [0060] In some wireless communications systems, for Type-II Port Selection codebook, only K (where K ≤ 2N 1 N 2 ) beamformed CSI-RS ports are utilized in downlink transmission, in order to reduce complexity. The. The KxN 3 codebook matrix per layer takes on the form * ^ = * H + I * ^,^ . Here, W2 follow the same structure as the conventional NR Rel. 15 Type-II Codebook, and are layer specific. * H + I is a Kx2L block-diagonal matrix with two identical diagonal blocks, e.g., * H + I = , J . ^ and E is an ^ × ^ matrix whose are vectors, as follows: J = ,3 ^^/^^ ^^/^^ ^^/^^ 1 KL^1MNLMN,^/^^ 3 1KL^1MNLMNP^,^/^^ … 3 1KL^1MNLMNPQR^,^/^^ /, where B is a standard unit vector with a 1 at the location. Here d PS is an RRC parameter which takes on the values {1,2,3,4} under the condition d PS ≤ min(K/2, L), whereas m PS takes on the values S0, … , T ^ ^ LMN U − 1W and is reported as part of the uplink CSI feedback overhead. W1 is common across all layers. [0061] For K=16, L=4 and dPS =1, the 8 possible realizations of E corresponding to mPS = {0,1,…,7} are as follows: Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 17 0 1 é 0 1 0 ù 0 1 0 ú ú ê ê 0 00 0 ú 0 00 0 , ú ê ê 0 00 0 ú ê 0 0 0 ú ê 0 00 0 û ë 1 00 [0062] When dPS =2, the 4 possible realizations of E corresponding to mPS ={0,1,2,3} are as follows: é 1 0 0 0 0 é 0 0 0 0 0 0 é 0 0 0 0 0 0 é 0 0 1 0 0 1 0 ù 0 0 ù 0 0 ù 0 0 0 1 ù ê ê 0 01 0 ú ú ê ê 1 00 0 ú ú ê ê 0 00 0 ú ú ê 0 00 0 ú 0 00 1 0 10 0 0 00 ê ú ê0 00 0 ú 0 0 0 0 0 , ê 0 01 0 ú , ê ú , ê ú . ê ú ê ú ê 1 00 0 ú ê 0 00 0 ú ê 0 00 0 ú ê 0 00 1 ú ê 0 10 0 ú 0 00 0 ê 0 ú ê 0 00 0 ú ê ú 0 00 ê 0 01 0 ú ê 1 00 0 ú ë 0 0 0 0 û ë 0 0 0 0 û ë 0 0 0 1 û ë 0 1 0 0 û [0063] When d PS =3, the 3 possible realizations of E corresponding of m PS ={0,1,2} are as follows: é 1 0 0 0 0 0 0 0 0 0 1 0 0 1 0 0 ù é0 0 0 0 ù é0 0 0 1 ù ê ê 0 01 0 ú ú ê ê 0 00 0 ú ê 0 00 0 ú 0 00 1 ú ê ú ê ú ê 1 00 0 ú 0 00 0 0 00 0 , 0 10 0 , ê ú . ê ú ê ú ê 0 00 0 ú ê 0 00 0 ú ê 0 01 0 0 00 0 ê 0 ú ê ú ê ú 0 00 0 00 1 ú ê 1 00 0 ú ë 0 0 0 0 û ë 0 0 0 0 û ë 0 1 0 0 û [0064] When d PS =4, the 2 possible realizations of E corresponding of m PS ={0,1} are as follows: é 1 0 0 0 0 0 0 10 0 ù é 0 0 0 0 0 0 ù ê ê 0 01 0 ú ê 0 00 0 ú 0 00 ú ê ú ê 1 0 0 0 0 0 00 0 ú , ê 1 00 0 ú . ê ú ê ú ê 0 00 0 0 1 0 0 ê ú ú ê ú 0 00 0 ê 0 01 0 ú ë 0 0 0 0 û ë 0 0 0 1 û Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 18 [0065] To summarize, m PS parametrizes of the first 1 in the first column of E, whereas d PS represents the row shift corresponding to different values of m PS . [0066] NR Rel. 15 Type-I codebook is the baseline codebook for NR, with a variety of configurations. The most common utility of Rel. 15 Type-I codebook is a special case of NR Rel. 15 Type-II codebook with L=1 for rank indicator (RI)=1,2, wherein a phase coupling value is reported for each sub-band, e.g., W2,l is 2xN3, with the first row equal to [1, 1, …, 1] and the second row equal to 23 4^_∅@ , … , 3 4^_∅9a:; <. Under specific configurations, ϕ0= ϕ1 …= ϕ, e.g., wideband reporting. are used for each pair of layers. NR Rel. 15 Type-I codebook can be depicted as a low-resolution version of NR Rel. 15 Type-II codebook with spatial beam selection per layer-pair and phase combining only. [0067] For NR Rel. 16 Type-II codebook, some wireless communications systems consider that a gNB is equipped with a two-dimensional (2D) antenna array with N1, N2 antenna ports per polarization placed horizontally and vertically and communication occurs over N 3 PMI subbands. A PMI subband consists of a set of resource blocks, each resource block consisting of a set of subcarriers. In such case, 2N1N2N3 CSI-RS ports are utilized to enable downlink channel estimation with high resolution for NR Rel. 16 Type-II codebook. In order to reduce the uplink feedback overhead, a DFT-based CSI compression of the spatial domain is applied to L dimensions per polarization, where L<N1N2. Similarly, additional compression in the frequency domain is applied, where each beam of the frequency-domain precoding vectors is transformed using an inverse DFT matrix to the delay domain, and the magnitude and phase values of a subset of the delay-domain coefficients are selected and fed back to the gNB as part of the CSI report. The 2N1N2xN3 codebook per layer takes on the form: ^ * ^ = * ^ * b ^,^ * c d , ^ , where W 1 is a 2N 1 N 2 x2L block-diagonal two identical diagonal blocks, e.g., * + = . /, and B is an N1N2xL matrix with columns drawn from a 2D oversampled DFT matrix, as follows: 567 567 ^ 9 :; ^ = 2 4 4 5 <, Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 19 operation. Note that O 1 , O 2 oversampling factors are considered for the 2D DFT matrix from which matrix B is drawn. Note that W1 is common across all layers. W f is an N 3 xM matrix (M<N 3 ) with columns selected from a critically- sampled size-N 3 DFT matrix, as follows: * c,^ = & e ^@ e ^; ⋯ e ^fg:; ), 0 ≤ B ≤ ^ h − 1, . columns of B are reported, along with the oversampling index taking on O 1 O 2 values. Similarly, for W f,l , only the indices of the M selected columns out of the predefined size-N3 DFT matrix are reported. In the sequel the indices of the M dimensions are referred as the selected Frequency Domain (FD) basis indices. Hence, L, M represent the equivalent spatial and frequency dimensions after compression, respectively. Finally, the 2LxM matrix * b ^ represents the linear combination coefficients (LCCs) of the spatial and frequency DFT-basis vectors. Both * b ^ , Wf are selected independent for different layers. Magnitude and phase values of an approximately β fraction of the 2LM available coefficients are reported to the gNB (β<1) as part of the CSI report. Coefficients with zero magnitude are indicated via a per- layer bitmap. Since all coefficients reported within a layer are normalized with respect to the coefficient with the largest magnitude (strongest coefficient), the relative value of that coefficient is set to unity, and no magnitude or phase information is explicitly reported for this coefficient. Only an indication of the index of the strongest coefficient per layer is reported. Hence, for a single-layer transmission, magnitude and phase values of a maximum of ⌈2βLM⌉-1 coefficients (along with the indices of selected L, M DFT vectors) are reported per layer, leading to significant reduction in CSI report size, compared with reporting 2N1N2xN3 -1 coefficients’ information. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 20 [0069] For Type-II Port Selection only K (where K ≤ 2N 1 N 2 ) beamformed CSI-RS ports are utilized in downlink transmission, in order to reduce complexity. The. The KxN 3 codebook matrix per layer takes on the form: * ^ = * H + I * b ^,^ * c d , ^ . [0070] Here, * b ^,^ and W f,l follow conventional NR Rel. 16 Type-II Codebook, are layer specific. The matrix * H + I is a Kx2L block-diagonal matrix with the same structure as that in the NR Rel. 15 Type-II Port Selection Codebook. [0071] In some wireless communications systems, Rel. 17 Type-II Port Selection codebook follows a similar structure as that of Rel. 15 and Rel. 16 port-selection codebooks, as follows: * ^ = * lllH + I * b ^,^*c d , ^ . [0072] However, unlike Rel. 15 selection codebooks, the port-selection matrix l * llH + I supports free selection of the K ports, or more precisely the K/2 ports per polarization out of the N1N2 CSI-RS ports per polarization, e.g., mlog ^ q^^^^ ^ /2 r s bits are used to identify the K/2 selected ports per polarization, wherein this across all layers. Here, * b ^,^ and Wf,l follow the same structure as the conventional NR Rel. 16 Type-II Codebook, M is limited to 1,2 only, with the network configuring a window of size N ={2,4} for M =2. Moreover, the bitmap is reported unless β=1 and the UE reports all the coefficients for a rank up to a value of two. [0073] For codebook reporting, a codebook report is partitioned into two parts based on the priority of information reported. Each part is encoded separately (Part 1 has a possibly higher code rate). Below we list the parameters for NR Rel. 16 Type-II codebook. [0074] Content of CSI report can include: Part 1: RI + channel quality indicator (CQI) + Total number of coefficients Part 2: SD basis indicator + FD basis indicator/layer + Bitmap/layer + Coefficient Amplitude info/layer + Coefficient Phase info/layer + Strongest coefficient indicator/layer [0075] Furthermore, Part 2 CSI can be decomposed into sub-parts each with different priority (higher priority information listed first). Such partitioning is required to allow dynamic reporting Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 21 size for codebook based on available resources the uplink phase. More details can be found in clause 5.2.3 of 3GPP technical specification (TS) 38.214. [0076] Also Type-II codebook is based on aperiodic CSI reporting, and only reported in physical uplink shared channel (PUSCH) via downlink control information (DCI) triggering (one exception). Type-I codebook can be based on periodic CSI reporting (physical uplink control channel (PUCCH)) or semi-persistent CSI reporting (PUSCH or PUCCH) or aperiodic reporting (PUSCH). [0077] For priority reporting for CSI report Part 2 multiple CSI reports may be transmitted with different priorities, as shown in Table: Table 1: Priority Reporting Levels for Part 2 CSI Priority 0: Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 22 Priority 4: Grou 2 CSI for CSI re ort 2 if confi ured as Priority 2 tuv − 1: , [0078] Note that the priority of the N Rep CSI reports can be based on the following: 1. A CSI report corresponding to one CSI reporting configuration for one cell may have higher priority compared with another CSI report corresponding to one other CSI reporting configuration for the same cell. 2. CSI reports intended to one cell may have higher priority compared with other CSI reports intended to another cell. 3. CSI reports may have higher priority based on the CSI report content, e.g., CSI reports carrying L1-reference signal received power (RSRP) information have higher priority. 4. CSI reports may have higher priority based on their type, e.g., whether the CSI report is aperiodic, semi-persistent or periodic, and whether the report is sent via PUSCH or PUCCH, may impact the priority of the CSI report. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 23 [0079] In light of that, CSI reports may be as follows, where CSI reports with lower IDs have higher priority Pri Bz{| ^}, , ~, ^^ = 2 ∙ ^ ^u^^^ ∙ ^ ^ ∙ } + ^ ^u^^^ ∙ ^ ^ ∙ + ^ ^ ∙ ~ + ^ s: CSI reporting configuration index, and Ms: Maximum number of CSI reporting configurations c: Cell index, and N cells : Number of serving cells k: 0 for CSI reports carrying L1-RSRP or L1- Signal-to-Interference-and-Noise Ratio (SINR), 1 otherwise y: 0 for aperiodic reports, 1 for semi-persistent reports on PUSCH, 2 for semi-persistent reports on PUCCH, 3 for periodic reports. [0080] For triggering aperiodic CSI reporting on PUSCH, a UE needs to report the needed CSI information for the network using the CSI framework in NR Release 15. The triggering mechanism between a report setting and a resource setting can be summarized in Error! Reference source not found. below. Table 2: Triggering mechanism between a report setting and a resource setting Periodic CSI SP CSI reporting AP CSI ng [0081] Moreover, ^ All associated Resource Settings for a CSI Report Setting need to have same time domain behaviour. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 24 ^ Periodic CSI-RS/ interference (IM) resource and CSI reports are considered to be present and active once configured by RRC. ^ Aperiodic and semi-persistent CSI-RS/ IM resources and CSI reports needs to be explicitly triggered or activated. ^ Aperiodic CSI-RS/ IM resources and aperiodic CSI reports, the triggering is done jointly by transmitting a DCI Format 0-1. ^ Semi-persistent CSI-RS/ IM resources and semi-persistent CSI reports are independently activated. [0082] FIG. 3 illustrates an aperiodic trigger state 300 defining a list of CSI report settings. For instance, for aperiodic CSI-RS/ IM resources and aperiodic CSI reports, the triggering is done jointly by transmitting a DCI Format 0-1. The DCI Format 0_1 contains a CSI request field (0 to 6 bits). A non-zero request field points to a so-called aperiodic trigger state configured by RRC, such as illustrated in FIG. 2. An aperiodic trigger state in turn is defined as a list of up to 16 aperiodic CSI Report Settings, identified by a CSI Report Setting identifier (ID) for which the UE calculates simultaneously CSI and transmits it on the scheduled PUSCH transmission. [0083] FIG. 4 illustrates an information element 400 pertaining to CSI reporting. For instance, when the CSI Report Setting is linked with aperiodic Resource Setting (e.g., including multiple Resource Sets), the aperiodic non-zero power (NZP) CSI-RS Resource Set for channel measurement, the aperiodic CSI-IM Resource Set (if used) and the aperiodic NZP CSI-RS Resource Set for IM (if used) to use for a given CSI Report Setting are also included in the aperiodic trigger state definition. For aperiodic NZP CSI-RS, the quasi co-located (QCL) source to use is also configured in the aperiodic trigger state. The UE considers that the resources used for the computation of the channel and interference can be processed with the same spatial filter e.g. quasi‐ co‐located with respect to “QCL‐TypeD.” [0084] FIG. 5 illustrates an information element 500 for RRC configuration for wireless resources. The information element 500, for instance, can configure NZP-CSI-RS/CSI-IM resources. The information element 400, for instance, illustrates RRC configuration (a) for NZP- CSI-RS Resource and (b) for CSI-IM-Resource. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 25 [0085] Table 3 presents types of uplink used for CSI reporting as a function of the CSI codebook type. Table 3: Uplink channels used for CSI reporting as a function of the CSI codebook type Periodic CSI SP CSI reporting AP CSI reporting reporting [0086] For aperiodic CSI reporting, PUSCH-based reports are divided into two CSI parts: CSI Part1 and CSI Part 2. The reason for this is that the size of CSI payload varies significantly, and therefore a worst-case uplink control information (UCI) payload size design would result in large overhead. [0087] CSI Part 1 has a fixed payload size (and can be decoded by the gNB without prior information) and contains the following: • RI (if reported), CRI (if reported) and CQI for the first codeword, • number of non-zero wideband amplitude coefficients per layer for Type II CSI feedback on PUSCH. [0088] FIG. 6 illustrates a scenario 600 for partial CSI omission for PUSCH-based CSI. The scenario 600, for example, illustrates reordering of CSI Part 2 across CSI reports. CSI Part 2 can have a variable payload size that can be derived from the CSI parameters in CSI Part 1 and contains Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 26 PMI and the CQI for the second codeword > 4. For example, if the aperiodic trigger state indicated by DCI format 0_1 defines 3 report settings x, y, and z, then the aperiodic CSI reporting for CSI part 2 can be ordered as illustrated in the scenario 600. [0089] As mentioned above, CSI reports are prioritized according to: 1. time-domain behavior and physical channel, where more dynamic reports are given precedence over less dynamic reports and PUSCH has precedence over PUCCH. 2. CSI content, where beam reports (e.g., L1-RSRP reporting) has priority over regular CSI reports. 3. the serving cell to which the CSI corresponds (in case of carrier aggregation (CA) operation). CSI corresponding to the PCell has priority over CSI corresponding to Scells. 4. the reportConfigID. [0090] Some proposals pertaining to wireless communications systems discuss using deep learning methods to efficiently compress and send CSI information to the gNB. For instance, one proposal suggests using a multilayer neural network to compress the input CSI and then instead of the original CSI, send the compressed information. Further, this proposal can be enhanced using a multiresolution encoder/decoder and which can reduce the Mean Square Error (MSE) between a desired and generated output. [0091] Another proposal presents a scheme where the compressed continuous representation is first quantized and then transmitted to the gNB side. In a related proposal, a vector quantization scheme is presented using neural networks where the prior is learnt from the data rather than being static. This proposal, for instance, has been used for compressed transmission of images. [0092] One way to train such machine learning models for CSI information is to select a UE from an environment (e.g., ^ ^ with reference to system 200) and collect training data associated to the selected UE. The network structure, hyperparameters of the model, codebook values, and neural network weights can be determined based on the collected data. After training a model, the parameters related to each part, e.g., the UE and the network entity portions, can be transmitted to a corresponding node if not already available at that node. For example, if the model is trained at a network entity, the information regarding the weights of the different model components, the Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 27 quantization codebook, and/or the number of levels can be transferred to the UE using appropriate signalling schemes. [0093] Such a trained model may exhibit acceptable performance for ^ ^ as data collected from the user ^ ^ is used for training. The trained model, however, might have less optimal performance for other UE such as if some of the statistics of the channel at a new node (e.g., ^ 4 ) are different from ^ ^ . Further, the structure of the model might need to be changed if a network parameter changes. For instance, a number of bits that can be used in the feedback can be different for ^ 4 and ^ ^ resulting in a different determination and/or selection of the values of ^ and ^. This scheme, however, may involve multiple models (e.g., one for each particular UE) which can be complex to store, manage, and assign to a new UE. [0094] An alternative is to combine the training data of a set of UE and construct a single model for the entire set of UE. Such a model, however, may have inferior performance as there might be users which have significant difference in UE channel statistics. Thus, training a single model with inputs having different statistics may result in a model with average and sub-optimal performance over different UE types. [0095] Accordingly, solutions are described in this disclosure for different CSI feedback mechanisms that provide CSI feedback based on AI models. For instance, this disclosure presents details corresponding to signalling of training data, parameters of AI models, and CSI feedback based on such models. To reduce an amount of CSI feedback, implementations construct a latent representation of ^, e.g., a low order latent representation. Implementations described herein, for example, use data-driven approaches to determine a correlation between the entries of ^ to reduce the amount of CSI feedback. [0096] FIGs. 7a and 7b illustrate respectively a UE subsystem 700a and a network subsystem 700b of a CSI system 700 that supports AI for CSI in accordance with aspects of the present disclosure. In at least one implementation the network subsystem 700b is implemented at a network entity 102. As further detailed below, the CSI system 700 includes two branches, a scalar quantization branch (e.g., the lower branch) and a quantization using codebook branch, e.g., the upper branch. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 28 [0097] According to one or more two latent representations of input data are generated. In at least one example, the input data is the channel matrix H and/or based on the channel matrix such as a function of the channel matrix, e.g., channel covariance matrix, eigen decomposition such as at least one eigen vectors, singular value decomposition (SVD) such as the at least one vector of the left and/or right singular vectors, etc. According to implementations the latent representations contain “real” numbers and thus it may not be practicable to send the latent representations directly using a finite number of feedback bits. [0098] Accordingly, at the lower branch (e.g., scalar quantization branch), the UE subsystem 700a quantizes real values of a latent representation and sends the quantized version to the network subsystem 700b, e.g., network entity such as gNB. In at least one example, the quantization occurring in the lower branch is based on a linear quantization with Q levels. At the upper branch (e.g., the quantization using codebook branch) the UE subsystem 700a compares the latent representation against codewords of a codebook and then instead of sending the actual latent representation, the UE subsystem 700a can transmit the ID(s) and/or index(s) of at least one codeword based on a measure of correlation or similarity of the indicated codeword(s) and the actual latent representation, such as the closest codeword(s), a weighted combination of a subset of the codewords, etc. Note that the codewords of the codebook are not fixed and can be learned during a training phase. [0099] Additionally, the various blocks of the network subsystem 700b can be trained to use the bits received from the UE subsystem 700a (e.g., feedback CSI bits such as those corresponding to the two latent representations) to generate a desired output. In at least some examples, a training objective is to have the output data (e.g., reconstructed data) as similar as possible to the input data. Alternatively or additionally other objective functions (e.g., loss functions) may be used for training as well. [0100] In the CSI system 700 different blocks of the system and associated procedures for the feedback CSI data can be generated at the UE subsystem 700a (e.g., a transmitter node) and then used by the network subsystem 700b (e.g., a receiver node) for reconstruction of the input data. Different aspects and operations of the system 700 are now described. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 29 [0101] In the UE subsystem 700a, input is input to a neural network 704. One example of the input data 702 is the ^ matrix that defined above. In implementations the input data 702 is a three-dimensional matrix representing a channel between Tx-Rx antenna pairs (^ × ^) over frequency bands, ^, for a UE. In at least some examples, the frequency bands may represent the channel per subcarrier, per every x subcarriers, per subcarrier group such as a PRB or sub-PRB or RBG (resource block group), etc. Further, the input data 702 can be a function of the H matrix, e.g., a vector corresponding to a singular vector that is associated with a largest singular value of the matrix H. [0102] The neural network 704 can be implemented as a multilayer neural network, for example using a convolutional neural network (CNN). In implementations the neural network 704 can be shared between both upper and lower branches of the UE subsystem 700a. The intermediate tensor output of neural network 704 ("Int_t_0") may be of size ~0 × ^0 × ^0. A neural network 706 (e.g., a multilayer neural network such as a CNN) receives output from the neural network 704 and generates output 708. The output 708, for instance, is a 3D intermediate tensor of size ~1 × ^1 × ^1 (namely “Int_t_1”), where ^1 represents, e.g., a number of filters at the last convolutional layer of the neural network 706 using CNN. [0103] In at least some implementations for each input sample (and based on the neural network 706 weights), there will be ~1 × ^1 tensors of size 1 × ^1 at the output 708. Parameters ~1, ^1, and ^1, for instance, are the hyperparameters that are determined during the training phase. [0104] According to one or more implementations the UE subsystem 700a sends a representation of the output 708 to the network subsystem 700b using a quantization codebook 710. The quantization codebook 710, for instance, is composed of ^ tensors (codewords) of size 1 × ^1. Each of these tensors have an ID or index which can be represented using log ^ ^ bits, e.g., since there are ^ different codewords. [0105] Further to the UE subsystem 700a, a mapper module 712 receives the output 708 and for each of its ~1 × ^1 tensors, the mapper module 712 generates at least one ID (between 0 to J) which shows the ID of the codeword (from the quantization codebook 710) which has a closest and/or largest correlation to the output 708. For instance, for the output 708, the mapper module 712 maps the input tensor of ~1 × ^1 × ^1 to ~1 × ^1 IDs each can be represented using log ^ ^ bits to Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 30 generate an output 714. Different metrics distance) can be used to compute the closeness between the vectors of the output 708 and the codebook 710 to generate the output 714. [0106] The UE subsystem 700a further includes a neural network 716 which can be implemented as a multilayer neural network, e.g., using CNN. The neural network 716 receives the output from the neural network 704 (e.g., the intermediate tensor output "Int_t_0") and generates an output 718. The output 718, for instance, represents a 3D intermediate tensor of size ~2 × ^2 × ^2 (namely “Int_t_2”), where ^2 is, e.g., a number of filters at a last convolutional layer of the neural network 716 realized using CNN. Further, the parameters ~2, ^2, and ^2 are the hyperparameters that are determined during the training phase. The output 718 is not necessarily of 3D shape and may optionally be 1D or 2D tensors such as depending on the structure of the neural network 716. [0107] To enable the UE subsystem 700a to send the output 718 and/or some representation thereof to the network subsystem 700b and to reduce the communication overhead, it may first pass the output 718 through a quantizer module 720, which in at least some implementations represents a scalar quantizer. In at least one example, the quantizer module 720 quantizes each value of the output 718 into 2 ^ levels, e.g., each quantized value can be represented using ^ bits. The value of ^ and the type of quantization used by the quantizer module 720 can be determined during the training phase. Thus, the quantizer module 720 receives the output 718 as input, and the quantizer module 720 generates an output 722. The output 722, for instance, represents a tensor of size ~2 × ^2 × ^2 where each entry takes only one of the 2 ^ possible values. [0108] Accordingly, the UE subsystem 700a transmits a representation of the outputs 714, 722 (e.g., encoded representations of the outputs 714, 722) to the network subsystem 700b via a feedback link 724. The outputs 714, 722 and/or representations thereof are sent (e.g., with a source and/or channel code and a modulation) to the network subsystem 700b e.g., with the feedback CSI information bits. [0109] According to implementations, the outputs 714, 722 can be sent to the network subsystem 700b using ~1 × ^1 × log ^ ^ + ~2 × ^2 × ^2 × ^ bits (information bits). For instance, ^~1, ^1^ are the number of latent vectors at the upper branch, ^ is the number of codewords in the quantization codebook 710 at the upper branch, ^~2, ^2, ^2^ show the size of the latent Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 31 representation in lower branch, and ^ is the of level used in the scalar quantizer in the lower branch. [0110] At the network subsystem 700b the gNB side receives via the feedback link 724 an input 725 and an input 726 which represent the output 714 and the output 722, respectively. The network subsystem 700b feeds the input 725 to a demapper module 728 (e.g., in the upper branch) and the input 726 to a neural network 730, e.g., in the lower branch. The demapper module 728 takes as input the received the ~1 × ^1 “IDs” in the input 725 and replaces and/or maps them to the corresponding codeword of size 1 × ^1 from a quantization codebook 732 which includes ^ tensors (codewords) of size 1 × ^1. The demapper module 728 outputs an output 734, which in at least one implementation represents a 3D tensor of size ~1 × ^1 × ^1, e.g., “Int_t_3”. The quantization codebook 732 may be same or different than the quantization codebook 710 of the UE subsystem 700a. [0111] The network subsystem 700b further includes a neural network 736 which can be implemented as a multilayer neural network, e.g., using CNN. The neural network 736 takes the output 734 as input and generates an output 738 (“Int_t_4”). The output 738, for instance, is a 3D tensor of size ~4 × ^4 × ^4. Further, parameters ~4, ^4, and f4 are the hyperparameters that are determined during the training phase. [0112] As mentioned above the neural network 730 takes the input 726 as input. Accordingly, the neural network 730 generates an output 740 (“Int_t_5”). The output 740, for instance, is a 3D tensor of size ~5 × ^5 × ^5. Parameters ~5, ^5, and f5 are the hyperparameters that are determined during the training phase. In one example, ~4 = ~5, and ^4 = ^5. [0113] To assist in concatenation of the outputs 738, 740, parameters ~5 and ^5 may be equal to ~4 and ^4, respectively. Considering these design parameters, ~ ^ and ^ ^ can be used as the first two dimensions of outputs 738, 740, e.g., output 738 can have the size of ~ ^ × ^ ^ × ^4 and output 740 can have the size ~ ^ × ^ ^ × ^5. Accordingly, a concatenator module 742 concatenates the outputs 738, 740 along the third dimension (e.g., filter dimension) and constructs “Int_t_6”. Thus, “Int_t_6” can be a 3D tensor of size ~ ^ × ^ ^ × ^^4 + ^5^. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 32 [0114] The network subsystem 700b a neural network 744. The neural network 744, for instance, is a multilayer neural network, such as implemented using CNN. The neural network 744 takes “Int_t_6” (output of the concatenator module 742) as input and generates output data 746. The output data 746, for example, represents a reconstructed data representation of the input data 702 previously input to the UE subsystem 700a. The output data 746 can be shared between both upper and lower branches of the network subsystem 700b. In at least some implementations, to enable reconstruction of the original input data 702, the size of the output data 746 is ^ × ^ × ^. [0115] The following section presents implementation details for the system 700. In this section, “UE” can refer to the UE subsystem 700a and “network,” “network entity,” and/or “gNB” can refer to the network subsystem 700b. [0116] Considerations regarding the network structure: a. In one example, the output of the neural network 716 is designed to be in the range [- 1,1]. For example, this can be enabled by applying an appropriate activation function (e.g., “tanh”) for the last layer of the neural network 716. b. Assuming an ideal feedback channel, input 725 and input 726 may be equal to output 714 and output 722, respectively. They could be different in cases of a non-ideal feedback channel, e.g., some elements of inputs 725, 726 are received with errors, omission of some elements of outputs 714, 722 in the feedback CSI, etc. Such effects may be modelled appropriately in the network structure of the system 700. c. The neural network structure of the different neural networks of the system 700 can be hyperparameters and can be determined during the training phase. Note that they can be fixed during the inference phase. d. The total available feedback rates can be partitioned between the data used for transmission of output 714 and output 722. For instance, on the selection of ^ ~1, ^1 ^ , ^~2, ^2^, the number of codewords in the quantization codebook 710, e.g., ^, and the levels of scalar quantization, e.g., ^. e. The system 700 can be scaled down to: Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 33 ^ Use codebook-based branch only: For this case, the lower branch of the UE subsystem 700a can be turned off or not used. In addition, neural network 730 and the concatenator module 742 can be removed from the network subsystem 700b. ^ Use scalar quantization branch only: For this case the upper branch of the UE subsystem 700a can be turned off or not used. In addition, the demapper module 728, the neural network 736, and the concatenator module 742 can be removed from the network subsystem 700b. The codebook 732 may optionally not be implemented and/or utilized. ^ In some examples, the network subsystem 700b (e.g., gNB) may indicate to the UE subsystem 700a (e.g., UE) to use at least one of the codebook-based quantization branch only, scalar quantization branch only, or both codebook- based quantization branch and scalar quantization branch. ^ In some examples, the UE subsystem 700a may determine to feedback the output of at least one of the codebook-based quantization branch only, scalar quantization branch only, or both codebook-based quantization branch and scalar quantization branch. Such determination may be based on the input data 702, e.g., channel matrix ^. The UE subsystem 700a may indicate to the network subsystem 700b an indication of such determination, e.g., the feedback CSI is based on codebook-based quantization branch only, scalar quantization branch only, or both codebook-based quantization branch and scalar quantization branch. f. A similar framework can be used when the input data 702 is not directly equal to the matrix ^. The input data 702, alternatively, could be an input data represented as a 3D matrix, e.g., a DFT transformed version of the channel matrix or a matrix representing the one/several eigen vectors and/or eigen value(s) of the channel matrix in different frequency bands. Alternatively or additionally, the input data 702 may correspond to a set of at least one precoding vector that is associated with a downlink transmission from a network node to the UE. As some other examples, H Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 34 could be of size NxMxT where third dimension represents values at different time symbols and/or time slots or could of size NxMxZ where the third dimension represents a composite time/frequency domain e.g., a stacked or concatenated frequency and time-domain vectors. g. The entries of ^ can be complex numbers and since most of the neural network methods work with real numbers, a transformation can be employed from a complex domain to the real domain. For example, the real and imaginary parts of the input data 702 (of size ^ × ^ × ^) can be separated and then concatenated together to generate an input data of size 2^ × ^ × ^ which only has real values. The concatenation can happen in other dimensions as well. In another example, the system 700 may virtually extend the channel matrix with their conjugate and then use inverse fast Fourier transform (IFFT) to transform the extended data. The results, for instance, will be real numbers and can be used with neural networks. h. Some of the tensors may have a reduced dimensionality. For instance, the second and/or third dimension 3D tensors described above may have value 1 reducing to 1D or 2D tensors. [0117] Considerations regarding the training and/or inference phases: a. The input data 702 can be collected at the UE subsystem 700a and then depending on where the model will be trained, it can be used at the UE subsystem 700a or transferred to the network subsystem 700b. b. The neural network weights are initialized randomly for training. The neural network weights can be changed during the training phase in a way that reduces the loss function. The neural network weights can be fixed during the inference time. c. The tensors of the quantization codebooks may not be fixed and they can be determined during the training procedure. They can also be fixed during the inference phase. d. It can be considered that the network subsystem 700b quantization codebook 732 is the same as the UE subsystem 700a quantization codebook 710. For instance, after Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 35 the model is trained, there may one quantization codebook which will be used by both subsystem 700a and subsystem 700b. For example, assuming that the complete model has been trained at the network subsystem 700b, the resulted quantization codebook can be transferred to the UE subsystem 700a along the other weights of the neural network blocks that are used for the UE subsystem 700a. If the training phase happens at the UE subsystem 700a, the quantization codebook 710 and the weights of the network subsystem 700b blocks will be transferred to the network subsystem 700b. [0118] Considerations regarding the network loss function: a. One example of the objective function is to minimize the mean square error between the input data 702 and the output data 746 (reconstructed data). b. One method to have an end-to-end differentiable loss function is to consider that input 725 is equal to the output 708 and input 726 is equal to output 718 in the backpropagation phase. [0119] Considerations regarding communication requirements: a. If the model trained at the network subsystem 700b, some mechanisms for exchange of some information between the UE subsystem 700a and the network subsystem 700b can be provided. ^ A method to send input data (e.g., channel measurements (or any of its desired transformation)) to the network subsystem 700b ^ A method to send the final neural network weights of the UE subsystem 700a side to the UE subsystem 700a ^ A method to send a quantization codebook (e.g., learned codebook) to the UE subsystem 700a ^ A method to send the number of quantization levels, Q, of the quantizer module 720 to the UE subsystem 700a Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 36 ^ A method to transmit 714 and output 722 (and/or representations thereof) to the network subsystem 700b b. If the model is trained at the UE subsystem 700a, some mechanisms can be provided for exchange of some information between the UE subsystem 700a and the network subsystem 700b. ^ A method to send the final neural network weights of the network subsystem 700b side to the network subsystem 700b. ^ A method to send a quantization codebook (e.g., learned codebook) to the network subsystem 700b. ^ A method to transmit output 714 and output 722 (and/or representations thereof) to the network subsystem 700b. [0120] It should be noted that in at least some implementations the UE subsystem 700a is to have enough computational resources to perform the training. Further, the UE subsystem 700a may have access to sufficient training samples of an environment to create a model with appropriate generalizations, e.g., in scenarios where the same model is to be used by different UE. In at least some examples, a UE performing training may be a high-performance UE and/or AI/ML model training source with capabilities for model training (e.g., sufficient computational and/or memory resources) and model transfer to a gNB (e.g., via Uu interface) and/or a UE, e.g., via sidelink channels. [0121] According to one or more implementations, CSI feedback information (also referred to here as H-CSI) can be determined according to a codebook configuration including codebook subset restriction, time-domain behavior (e.g., periodic, semi-persistent, aperiodic), and measurement restriction configurations (e.g., restrictions on which RS symbols and/or slots can be used to determine the input data), e.g., for the upper branch of the system 700. a. The lower branch of the system 700 may also be independently associated with a different time-domain behavior and/or measurement restriction configuration than that of the upper branch of the system 700. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 37 b. In at least one example, a UE a codebook subset parameter or a rank parameter configuration, and a CSI report is tailored for a subset of permitted codewords or ranks. The UE determines if there is a trained neural network model available corresponding to the configured codebook subset/rank parameter. The UE can indicate to the network whether the model is available. A CSI report associated with the parameter can be provided if the corresponding neural network model is available after a time ‘T’ from at least one of (a) a CSI trigger or (b) after reception of the restriction configuration, or (c) after a transmission of an acknowledgment in response to reception of the parameter configuration. The time ‘T’ can depend on the configured parameter (e.g., for each available rank parameter, the network can configure a corresponding ‘T’ value or the UE can report a ‘T’ value via UE capability signalling to the network). [0122] In implementations, an RRC message/ DCI/MAC-control element (CE) associated with an H-CSI report (e.g., DCI/MAC-CE triggering the H-CSI report) can indicate one or more of whether and which of the parameters corresponding to the lower branch, upper branch, neural network weights, codebook, or a combination thereof are to be signaled. Alternatively or additionally, the UE in the corresponding H-CSI report can indicate such information. [0123] In implementations, an H-CSI report can include one or more subband H-CSI reports and a wideband H-CSI report and: a. The subband size associated with each subband H-CSI report can be reported by the UE and can be determined based on the neural networks. b. A subband/wideband H-CSI report can be derived based on the lower branch, upper branch, or the combination of the two branches, and the wideband H-CSI report can be independently configured/indicated to be based on the lower branch/upper branch or the combination of the two branches. c. In at least some examples, the wideband H-CSI report may be based on input data that has sparser frequency resolution than that used for subband H-CSI report e.g., input data based on a channel value every x PRBs or PRB-Groups, x=4 for wideband H-CSI report and x=1 for subband H-CSI report. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 38 [0124] In implementations, the network the UE to use only upper branch, only lower branch, or use both branches. [0125] In implementations, a H-CSI report can be provided by the UE (or H-CSI report is valid) if the H matrix is computed (e.g., based on the associated scheduled data) over a number of resource elements (RE) larger than a threshold. Further, a subband H-CSI report can be provided/valid if at least a wideband H-CSI report associated with the subband H-CSI report is computed over a larger than a threshold number of RE. [0126] In implementations, the parameters that can be signaled include at least one of: (i) an output of a scalar quantization branch (e.g., sent in an uplink direction), (ii) an output of a quantization codebook branch (e.g., sent in an uplink direction), (iii) a codebook comprising the codewords corresponding to the quantization codebook branch (e.g., sent in a downlink direction, such as when training occurs at network side), (iv) parameters corresponding to neural network nodes for each of the scalar quantization branch and the quantization codebook branch (e.g., sent in a downlink direction, such as when training occurs at network side), and (v) parameters related to the blocks shared between the two branches of the system 700. [0127] In at least one implementation, the parameters corresponding to the codebook and neural network parameters are computed at the network and signaled to the UE via at least one of: a. Downlink control information (DCI) signaling with a DCI format that carries information corresponding to CSI configuration; b. Higher-layer signaling, e.g., MAC-CE signaling or RRC signaling; or c. Combinations thereof. [0128] In at least one implementation, the parameters corresponding to the codebook and neural network parameters are computed at the UE and signaled to the network via at least one of: a. A part of a CSI report, where the CSI report comprises at least one part; b. A part of an AI/ML/neural network-based report; or c. Combinations thereof. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 39 [0129] In at least one implementation, the configures the UE to report at least one of the quantization codebook and the neural network parameters based on a report quantity included as part of a codebook configuration. [0130] In at least one implementation, the parameters corresponding to the scalar quantization branch are reported in a first CSI report part, where the first CSI report part is prior to a second CSI report part corresponding to the codebook quantization branch and the parameters related to the shared part which are reported subsequently. [0131] In at least one implementation, a UE is instructed to iteratively use the upper branch of the UE subsystem 700a for the first ^ ≥ 0 time slots, then use the lower branch of the UE subsystem 700a for the next ^ ≥ 0 time slots, and then use a model with both upper and lower branches for subsequent h ≥ 0 time slots. [0132] In at least one implementation, the output of at least one of the scalar quantization branch and the codebook quantization branch corresponds to PMI, CQI, and/or a combination thereof. [0133] Various implementations described herein provide for training of a two-sided model. For instance, consider that there are ^ users in the network and ^ ^ is a set of training data collected from ^ ^ . The term ℳ can be used to refer to a general two-sided CSI feedback model and ℳ u and ℳ L to refer to the encoder (e.g., UE part) and decoder part (e.g., gNB part) of the model, [0134] According to one or more implementations, separate models can be used for different user types and/or different network requirements. For instance, consider that the ^ users can be grouped into ^ types (^ ≤ ^), ^ ^ , ^ = ^1, … , ^^ , where the users in each type have similar statistics for their ^ ^ . With this assumption, we can combine training data of users of one type and construct the training data for that type, e.g., ^&^) = ^∪ ^ ^ , ∀ | ^ ^ ∈ ^ ^ ^. [0135] In such implementations, instead of a single model, there can be up to ^ different versions of ℳ (for example one for each type), where model ℳ&^), ^ = ^1,2, … , ^^ has been trained using the ^&^) training dataset. Since users of each type may have similar statistics and the Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 40 model can be trained using the same type of the performance of model ℳ & ^ ) can be suitable for UE of type ^. [0136] Other than the user type, ℳ can be changed if other parameters of the network change. One important network parameter is the number of feedback bits. With reference to the system 700 introduced above, a change of a number of feedback bits may result in change in values of ^, ^ and may also change the structure of the neural network blocks, e.g., values of ~ ^ , ^ ^ , ^ ^ , ~ ^ , ^ ^ , ^ ^ . Similar to scenarios for different user types, to accommodate different scenarios, an example implementation is to train and save different ℳs for different requirements, e.g., different numbers of feedback bits. [0137] For implementations involving multiple models, implementations can be provided to store different models and select and/or use an appropriate model based on the type of the UE, e.g., when a UE connects to a network and/or is configured for CSI feedback based on a two-sided model. For example, different models can be stored at a network entity. Accordingly, when a new UE connects to the network and/or is configured for CSI feedback based on two-sided model, the network entity can decide which model is to be used for that UE. For instance, channel measurements can be collected from that UE and checked against different models to see which model is a best fit for the UE. The network entity can then send the UE one or more portions of the model, e.g., ℳ u &^) for the UE, such as based on the system 700, which may include transmission of a quantization codebook and quantization level. In at least some implementations, a model selection neural network may be used to select a model for CSI feedback based on UE channel measurements and/or a representation thereof. For example, the model selection neural network may be transmitted to the UE and the model selection procedure is performed at the UE. Alternatively or additionally, the model selection procedure is performed at the network entity and/or another network node based on UE channel measurement feedback from the UE. [0138] In at least some implementations, different models can be preloaded (e.g., via higher layer signaling or configuration such as RRC) to a UE and a network entity informs the UE which model (e.g., via a model index) is to be used. Such implementations can reduce signalling when a model switch is implemented. Further, model parameters can be transferred to the UE beforehand, such as when the UE wakes-up, connects, is configured for CSI feedback based on model(s), and/or in conjunction with manufacture and/or initial configuration and deployment of the UE. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 41 [0139] In at least some implementations UE of an environment have similar ^ statistics, a network entity may have a single model available. Based on changes in the UE and/or environment, the network entity can decide that a current model is not a good fit. In such scenarios, an update procedure can be started to retrain the model. The retraining may be performed at a UE, network entity, and/or other node. Parameters of the updated/retrained model can subsequently be sent to a corresponding network entity and UE, such as if not already available at that node. Such implementations can reduce the need to train and save multiple models and also may avoid a model selection procedure. [0140] In at least some implementations as part of a model update procedure, initial weights of the model may be random numbers (e.g., based on a uniform or gaussian distribution), weights of the current model, weights of a model trained using a meta learning scheme, etc. The weights of the meta learning scheme, for instance, can be determined by training the model using a meta learning approach such as model-agnostic meta-learning (MAML) and using the data collected using multiple user types. [0141] According to implementations, the discussion above can be applied when ^ = 1 and/or for a single set of network requirements. Further, if there are multiple UE types and multiple network requirements, multiple models can be constructed based on both UE type and network requirement. [0142] Implementations described in this disclosure also enable a single model for different numbers of feedback bits. For instance, implementations enable training, deployment, and operation of two sided models when a number of feedback bits is not constant. For simplicity of explanation, consider that there may be one user type. However, for multiple user type scenarios, the described approaches can be combined with implementations described above. [0143] For example, considering the system 700, an amount of information that a UE (e.g., the UE subsystem 700a) sends to a network entity (e.g., the network subsystem 700b, e.g., gNB) can be based on one or more of a) size of the latent representation that can be quantized using the quantizer module 720 (e.g. the output 718), b) a number of bits used for quantizing each value (e.g., log ^ ^ ^ ^ of the quantizer module 720), c) a number of vectors that will be quantized using the quantization codebook 710 ( e.g. the output 714), d) the number of bits used to select a quantization codebook Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 42 codeword, e.g., log ^ ^ ^ ^ of the mapper module For instance, in the current example, a total number of bits that can be sent to the network subsystem 700b can be ^ ^ × ~ ^ × log ^ ^ ^ ^ + ^ ^ × ~ ^ × ^ ^ × log ^ ^ ^ ^ . [0144] To enable different numbers of feedback bits to be supported, the system 700 described above can be utilized and based on a number of feedback bits, create different models by adjusting ^ ^ , ~ ^ , ^, ^ ^ , ~ ^ , ^ ^ , and ^ and alternatively or additionally, the neural networks 704, 706, 716. [0145] In at least some implementations that utilize a single model and support multiple feedback rates, ^ ^ , ~ ^ , ^, ^ ^ , ~ ^ , and ^ ^ can be fixed the quantization levels changed, e.g., ^. Considering this, a network can be implemented which uses ^ ^ × ~ ^ × log ^ ^ ^ ^ + ^ ^ × ~ ^ × ^ ^ × ℓ bits where ℓ is an integer number, ℓ = ^0,1, … ^ and represents the number of bits used for sending each value after value quantization. The value of ℓ may be configured, provided, and/or indicated to a UE such as via CSI report configuration, RRC, MAC-CE, DCI (e.g., for semi-persistent and/or aperiodic CSI reporting), etc. Further, the value of ℓ may be selected, reported, and/or determined by the UE based on a number of resources for CSI feedback, such as a number of REs and/or a coding rate, where a CSI report code rate is less or equal to a threshold (e.g., configured by the higher layers), maximum CSI UCI information bits and/or payload size, etc. In implementations where the value of ℓ selected and/or reported by the UE, ℓ may be included in the Part 1 of a CSI report for a CSI report comprising multiple (e.g., two) parts. [0146] Various implementations enable training, selection, and operation of models separately for each ℓ. Further, if a model is changed only using ℓ , the models can have a property where the structure of the neural networks (e.g., of the system 700) can remain the same between models corresponding to different feedback rates. Therefore, a single ℳ can be utilized (e.g., corresponding to a fixed neural network structure and quantization codebook) and the two-sided model can be trained such that it supports different values of ℓ. For instance, a same training data set can be applied with the value of ℓ changing from one sample to another during the training phase. Using such implementations, the neural networks in the ℳ u and ℳ L can be trained even though a quantization level might change from time to time. some implementations the network entity can implement the ℳ L part and ℳ u can be implemented by the UE. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 43 [0147] Such implementations may reduce the need to update the ℳ u or ℳ L after each change in the number of feedback bits and a network entity may notify a UE of a new ^ and other weights of the neural network blocks can remain same, e.g., unchanged. This may signaling overhead and latency due to model selection. [0148] In at least some implementations, different values of ℓ can be provided to support different feedback rates by rounding, truncating, and/or dropping one or more least significant bits (LSB) of a reference ^-level quantizer of the latent representation, e.g., on the lower branch of the system 700. Further, in at least some implementations, the upper branch of the system 700 may be not present, and the system 700 utilizes the scalar quantization. [0149] In at least some implementations, the neural networks of the system 700 can be represented as respective blocks and at least one block can be further decomposed into a plurality of sub-blocks. In one example, the neural network 716 is decomposed into blocks B31, B32, .., B3K and neural network 730 is decomposed into B 51 , B 52 , .., B 5K , such that sub-block B 3k can be coupled with sub-block B 5k . In such implementations, a first of two models may comprise a first subset of sub-blocks, e.g., B31, B33 under neural network 716, B51, B53 under neural network 730, and a second of the two models comprises a second subset of sub-blocks, e.g., B31, B32 under neural network 716, and B 51 , B 52 under neural network 730. [0150] In at least some implementations, a UE is configured (e.g., using an associated neural network model configuration) with a plurality of feedback rates and/or quantization granularities corresponding to a neural network. In at least one example, a minimum number of bits and a maximum number of bits is configured, and the range in between (with potential skipping) is supported. [0151] In at least some implementations, a UE determines a feedback rate based on a corresponding CSI report configuration, e.g., based on one or more of PUCCH-CSI-Resources in CSI-ReportConfig. The UE may not be configured with a list of PUCCH resources in the corresponding CSI report configuration that is not associated with and/or cannot accommodate the configured feedback rates. In an alternative or additional implementation, the UE is configured with a particular feedback rate (e.g., in the corresponding CSI report configuration), and determines a Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 44 PUCCH resource from a list of PUCCH to convey the CSI report based on the CSI report payload size. [0152] In at least some implementations, a UE indicates a feedback rate used along with the CSI report, or the network entity infers the feedback rate based on the PUCCH resource carrying the CSI report. In the latter case, another CSI report (e.g., L1-RSRP or RI) may be indicated to not to be multiplexed with the CSI report corresponding to latent representation of the channel. [0153] In at least some implementations, a UE may not be configured to provide an aperiodic CSI report sooner than: ^ a 1 st specified time from a DCI triggering a CSI report for a 1 st feedback rate, and ^ a 2 nd specified time from a DCI triggering CSI report for a 2 nd feedback rate. [0154] In at least some implementations, if a feedback payload part of a quantized representation of a latent representation (^ ^ × ~ ^ × ^ ^ × ℓ ) with a first determined ‘l’ value (e.g., ‘l is ’RRC/MAC-CE/DCI indicated or determined according to a CSI report configuration) corresponds to a first PUCCH resource for transmitting the CSI report that is not available (e.g., due to collision of the PUCCH with a downlink symbol): ^ The UE (e.g., if configured and/or indicated, and/or has sufficient time) provides the feedback payload based on a second determined ‘l’ value, o wherein the second ‘l’ value is smaller than the first ‘l’ value if a second PUCCH resource is available that can carry the feedback payload. ^ Wherein the availability is conditioned on satisfying some timelines, ^ e.g., not earlier than a first threshold or not later than a second threshold, wherein the thresholds are defined with respect to and/or relative to a DCI triggering the CSI report (e.g., a last symbol of the DCI) and/or with respect to the first PUCCH resource, e.g., a first symbol of the first PUCCH resource. [0155] Implementations described herein also provide solutions to address CSI report collision. For instance, if two CSI reports collide when a UE is configured to transmit two colliding CSI Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 45 reports, the UE based on indication and/or can change a quantization rate (‘l’) of one of the reports and transmit both reports instead of dropping one of the two reports. For instance, a 1st CSI report is associated with a quantized representation of one latent representation, and a 2 nd CSI report is associated with one or more of CQI, PMI, CSI-RS resource indicator (CRI), synchronization signal (SS)/physical broadcast channel (PBCH) Block Resource indicator (SSBRI), layer indicator (LI), rank indicator (RI), L1-RSRP, L1-SINR or Capability[Set]Index. Thus, the UE can lower ‘l’ for the 1 st report and send both 1 st and 2 nd CSI reports. [0156] In at least some implementations a two-sided model is provided with flexible UE portions. For instance, as discussed above, multiple models can be provided for each group and/or type of UE, and/or for each network parameter. To improve the performance of such implementations, the following discussion presents an approach to enable fine tuning of models for each UE without undue complexity. For instance, a network entity uses a same ℳ L for UE of a same category (e.g., with a single set of model properties) and the UE are able to use different encoding schemes, which may increase performance. For simplicity of explanation in the following discussion, consider that the network requirements (e.g., number of allowed feedback bits) are fixed. However, if a number of allowed feedback bits are dynamic (e.g., changing), the following implementations can be combined with implementations described above. [0157] Thus, according to at least some implementations that enable flexible UE portions: 1- Consider that a pretrained two-sided model (e.g., ℳ u and ℳ L ) is already selected (e.g., if more than one pretrained two-sided model exist) for a particular UE. 2- The decoding part of the model, ℳ L , is then deployed at the network entity, e.g., it can be transmitted to the network entity if training happens at a different node. The encoding part of the model, e.g., ℳ u , is transferred to the UE by the network entity and/or other network node. 3- In at least some implementations, a network node (e.g., gNB) can send ℳ L to the UE. 4- The UE can use the ℳ u to construct the feedback bits based on input channel data. Further, a user can be enabled to fine tune ℳ u . Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 46 5- For fine tuning the user, e.g., ^ 4 can and ℳ L to construct a version of the complete model locally. ^ 4 can also collect channel as local training data ^ ^ 4 . In at least some implementations, the UE can use a part of the original training data which has similar statistics to its channel condition. 6- ^ 4 then trains the local model using ^ ^ 4 . In at least one implementation as part of ^ 4 training the local model using ^ ^ 4 , ^ 4 is not permitted to change a parameter of the ℳ L , e.g., weights and/or quantization codebook of ℳ L are fixed. The fine-tuned parameters of the encoding part (e.g., ℳ u ^ at user ^ can be by ℳ ^ u 4 . 7- After completion of the training step, ^ 4 can deploy and use ℳ ^ u 4 (e.g., as fined tuned for ^ 4 ) as a UE part of the model, e.g., the UE subsystem 700a. [0158] Note that in at least some implementations, each UE has its own encoding neural network while a network entity may have a single decoding neural network ℳ L . Thus, the existence of several models and performance of model selection may be avoided where each UE adapts the encoding part locally. [0159] Note that in at least some implementations, a UE may change some parts of ℳ L as well during fine tuning. In such implementations, the UE can feedback the updated parts to a network entity. [0160] In at least some implementations, such as when a UE does not have sufficient computational capability for training, the UE can send ^ ^ 4 to a training node (e.g., a network entity and/or other node) and training can be performed at the training node. The training node, for example, can use training data which has similar statistics to its channel condition. This information can be transferred to the training node if it is not already available. After completion of the training phase, the updated ℳ u can be sent from the training node to the UE. [0161] In at least some implementations, a UE may also train a local model considering that it can also modify the quantization codebook. In such scenarios, after completion of training, the UE may send the modified quantization codebook to a network entity while other parts of ℳ L may remain fixed. Each UE, for instance, has its own encoder ℳ ^ 4 , and u network entities use a same Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 47 neural network structure for the group of UE. instance, the demapper module 728 uses the quantization codebook associated with each UE. [0162] In at least some implementations of such scenarios, a single model at the network entity can be utilized with multiple quantization codebooks. For instance, model selection may be avoided as each UE can use its own ℳ ^ u 4 and quantization codebook. [0163] Considering that a UE can report an updated quantization codebook to a network entity, at least one of the following may be considered: ^ The updated quantization codebook is in a form of additional codewords to be appended to an original quantization codebook; ^ The updated quantization codebook is in a form of additional bits appended to each codeword of the original quantization codebook; and/or ^ The updated quantization codebook is reported in response to a network configuration parameter that configures the UE to report an updated quantization codebook. [0164] In at least some implementations, a single instance of the upper branch or the lower branch of the system 700 might be not present, and a variation on the system 700 may operate using an instance of the vector/codebook quantization or scalar quantization. [0165] In one or more implementations, CSI feedback is based on AI/ML model selection. Assume a channel between a UE 104 and a network entity 102 with P channel paths (index ^ = 0, … , ^ − 1) that occupies NSB frequency bands (index ' = 0, … , ^ {^ − 1), where the network entity (e.g., gNB) is equipped with K antennas (index = 0, … , ^ − 1). The channel at a time index δ can then be represented as follows ¨R^ £ L £ where gk,p spacing, τp refers to delay of path p, F c refers to carrier Frequency, c refers to speed of light, d refers to antenna spacing at the network entity, θp refers to angular spatial displacement at the network entity antenna Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 48 array corresponding to path p, δ refers to time v refers to relative speed between the network entity and the UE, and Φ p refers to angle between the moving direction & the signal incidence direction of path p. [0166] The channel above is parametrized by three dimensions: frequency, spatial, and temporal dimensions. While for most scenarios/use cases the channel is assumed to be fixed for a long- enough interval of time to pursue CSI measurement, reporting and signal precoding via NR-based linear precoding techniques within the channel coherence time, this assumption may be revisited for other precoding techniques, e.g., AI/ML-based techniques that require sharing the AI model parameters, for which the overhead can be tremendously large and hence would need to be carried out over a large number of slots. Note that a change in the channel behavior may be associated with a change of the UE orientation, a change of the UE line-of-site (LoS) condition, or a combination thereof. More specifically, assume a network with gNB-centric AI/ML training/modeling. While UE within a coverage area of the network entity (e.g., gNB) may correspond to infinitely many instantaneous channel coefficient values, these channel coefficient values can be categorized under a finite number of channel distributions, e.g., based on region of UE location, indoor/outdoor UE status, LoS/NLoS UE status, or some combination thereof. Due to the large variation of the channel distributions a common model that is trained using a mixture of training data based on the aforementioned distribution may not be generalizable enough to provide high-resolution precoding for all distributions. In light of that, multiple training data sets that can be used to train multiple training models are needed for a higher resolution precoding. [0167] In the discussions herein, a CSI framework that supports multiple AI/ML models is discussed, where each AI/ML model corresponds to a distinct channel distribution. In order to support such a framework, a training data set may be partitioned into multiple training data sets, where each training data set is used to train a distinct AI/ML model. Techniques of classifying and possibly sharing the training data set between the UE and the network, as well as selecting the AI/ML model and sharing the AI/ML model parameters and selections between the UE and the network will be discussed in further detail below. Furthermore, a second stage precoder adjustment is discussed, where the second stage precoder adjustment enables switching ON/OFF and/or scaling the amplitude corresponding to a subset of CSI-RS ports, a subset of indices of a transformed spatial domain, a transformed frequency domain, a transformed time domain, or a combination thereof. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 49 Several implementations that describe the CSI framework are described below. One or more elements or features from one or more of the described implementations may be combined. [0168] A description of the AI/ML model for CSI measurement and reporting follows. For AI/ML-based CSI frameworks, multiple alternatives exist for the outline of the AI/ML algorithm functionality, as follows. In one or more implementations, the AI/ML model is trained at the UE. This alternative may appear reasonable since the UE is the node that can seamlessly collect training data for CSI acquisition using downlink (DL) pilot signals, e.g., CSI-RSs for channel measurement, however, the AI/ML model may be re-trained whenever the environment changes, e.g., change of the UE location or orientation and every training instance requires significant memory and computational complexity requirements. [0169] Additionally or alternatively, the AI/ML model is trained at the network entity. One advantage of this approach is that the network has significantly more power and computational capabilities compared with a UE, and hence can manage training moderately complex AI/ML models, as well as store large amounts of training data. Moreover, since a network node is mostly assumed to be fixed, its coverage area is expected to be the same and hence a single AI/ML model can be applicable to UE within a specific region of the cell for a reasonable period of time. One challenge with this approach is related to obtaining the training data at the network node, especially for FDD systems in which the uplink/downlink (UL/DL) channel reciprocity may not hold. Note that the overhead corresponding to feeding back the training data from the UE to the network may be considered as one of the metrics when assessing the efficiency of an AI/ML algorithm. [0170] Note that a series of model re-training events and their associated measurement configuration can be defined. For instance, in case of UE-based AI/ML model training, the UE upon determining a significant change in one or more of parameters of a channel distribution over a measurement window can determine that a channel distribution change has occurred, and hence triggering model re-training. Additionally or alternatively, the network node upon determining a significant change in one or more of parameters of a channel distribution over a measurement window can determine that a channel distribution change has occurred, and hence triggering model re-training Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 50 [0171] In some discussions it is assumed model is trained at the network due to the advantages corresponding to memory, computation, and cell-centric characteristics of the network- based AI/ML model computation. The challenge corresponding to obtaining the training data corresponding to the DL channel at the network side is discussed below. [0172] Assuming the AI/ML model is trained at the network, a few aspects of DL training data acquisition at the network side to enable efficient AI/ML modeling are discussed below. In one or more implementations, in order to maintain the robustness of the AI/ML model with respect to channel variations, DL training data may be continuously fed back to the network to keep up with changes in the environment, e.g., traffic, weather, and mobile scatterers. Note that this may not necessarily correspond to online learning; even for an offline learning algorithm a framework for obtaining new training data corresponding to channel variations may be characterized. In one example, the UE is configured with (pseudo) periodic training periods, where the UE can feedback information which could help the network node adjust/update the AI/ML model parameters if needed. The feedback can be sent using configured grant PUSCH transmissions. [0173] In one or more implementations, based on the current codebook-based DL CSI feedback schemes in NR, the CSI is compressed in at least one of the spatial domain, or the frequency domain, or both. One approach would be using the codebook-based CSI feedback, e.g., Type-I and/or Type-II codebooks for obtaining the training data. One disadvantage of this approach is that the training data would comprise CSI feedback that is already compressed via conventional approaches, which would have detrimental effect on the AI/ML model inference accuracy. For instance, if the AI/ML model compares the output of the AI/ML model with the channel corresponding to the CSI feedback to assess its own inference accuracy, this assessment would not be precise since it is based on H’, an estimate of the channel based on a pre-defined compression, rather than H, a digitally quantized channel without further compression in spatial domain, or frequency domain. On the other hand, if the UE feeds back the training data corresponding to the DL CSI feedback without compression over spatial and/or frequency dimensions, the feedback overhead of the training data would be significant, which would beat the purpose of using the AI/ML model, which is mainly to reduce the overall CSI feedback overhead. For example, numerically an AI/ML-based CSI feedback aims at minimizing the following metric ®­ ^ − ­® Attorney Docket No. SMM920220072-WO- Lenovo Docket No. SMM920220072-WO-PCT 51 where H represents a digital-domain of the channel matrix. On the other hand, a compressed channel H’, which represents the recovered channel after codebook-based transformation, would yield the following optimization metric m ­^ in ®­ ^ − ­′® Since ­ ≠ ­′, the output of both different channel estimates. [0174] For DL CSI acquisition in NR, whether the network operates in FDD mode or TDD mode, it is unlikely that AI/ML would fully replace RS-based CSI feedback for high-resolution precoding design, since some channel parameters may vary from one time instant to another, without strong correlation across the two time instants, e.g., initial random phases of the channel. Given that, AI/ML-based CSI framework can be envisioned as a technique for further reducing the CSI feedback overhead compared with conventional methods, e.g., reduce the number of dominant spatial-domain basis indices, frequency/delay-domain basis indices, and time/Doppler-domain basis indices, after spatial domain transformation, frequency-domain transformation, and time-domain transformation, respectively. While current CSI feedback frameworks already provide CSI feedback overhead reduction via exploiting such transformations, the CSI dimensionality can be further reduced if a wider range of transformation techniques are pre-configured, where a different transformation may be selected for a given UE based on variations of the channel. [0175] CSI feedback corresponding to a selection from multiple AI/ML models is discussed below. In the following implementations, a CSI feedback framework is proposed that enables generalized codebook reporting corresponding to a variety of channel conditions, e.g., UE location, indoor/outdoor UE status, LoS/NLoS UE status, or some combination thereof. Several implementations that describe the proposed CSI framework are provided below. One or more elements or features from one or more of the described implementations may be combined. [0176] In one or more implementations, training data partitioned based on different channel conditions. A training dataset corresponding to CSI is partitioned into multiple subsets of training dataset, where each subset of the multiple subsets of the training dataset corresponds to a distinct channel condition. Several examples of partitioning the training data set are discussed in the following. In one example, the training dataset is partitioned based on a ratio of a channel gain of a strongest spatial domain index, e.g., CSI-RS port, to that of a second strongest spatial domain index. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 52 E.g., if the ratio is at or above a threshold the training dataset is in one subset, and if the ratio is below the threshold value, then the training dataset is in another subset. Additionally or alternatively, the training dataset may be partitioned based on a ratio of a channel gain of the strongest spatial domain index to that of a remainder of the spatial domain indices. [0177] In another example, the training dataset is partitioned based on a ratio of a channel gain of a strongest frequency domain index, to that of a second strongest frequency domain index. Additionally or alternatively, the training dataset may be partitioned based on a ratio of a channel gain of the strongest frequency domain index to that of a remainder of the frequency domain indices. [0178] In another example, the training dataset is partitioned based on a location estimate of a UE, where a first UE whose estimated location is within a first region is partitioned into a first subset of the multiple subsets of the training dataset, and a second UE whose estimated location is within a second region is partitioned into a second subset of the multiple subsets of the training dataset. [0179] In another example, the training dataset is partitioned based on a CRI value corresponding to the UE, where a first UE whose corresponding CRI value is equal to a first value is partitioned into a first subset of the multiple subsets of the training dataset, and a second UE whose corresponding CRI value is equal to a second value is partitioned into a second subset of the multiple subsets of the training dataset. [0180] In another example, a first subset of the multiple subsets at least contains a first number of elements (‘n1’), and a second subset of the multiple subsets at least contains a second number of elements (‘n2’) where the first number and the second number are configured or determined. Note that ‘n1’ and ‘n2’ may be related; for instance, ± < ^^ ^ ^ < ². Additionally or alternatively, in case the UE cannot provide a given number of training samples, e.g., ‘n1’, within a certain time window, the UE indicates such an event to the network. [0181] In one or more implementations, the UE feeds back CSI to the network, where the fed back CSI corresponds to an AI/ML-based training dataset, and the CSI comprises a parameter corresponding to an indicator of a subset of the dataset from the multiple subsets of datasets. In one example, the parameter is included as part of a CSI report that is fed back to the network over one Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 53 of PUSCH, PUCCH, where reporting the is triggered, e.g., via DCI triggering a CSI report. [0182] In another example, the indicator value is computed based on one of a fixed or higher- layer configured formula that is indicated by the network to enable a classification of the CSI to one of the multiple subsets of the training dataset. [0183] In another example, an AI/ML model training or retraining event is triggered due to a change in a channel condition, and the UE notifies the network about the event via a first notification. The network or UE determines the subset of the dataset from the multiple subsets of datasets based on the first notification. The subset of the dataset is valid for a pre-determined duration of time or until another notification is received. [0184] In another example, the CSI reporting setting associated with the CSI report includes a field which indicates if the CSI report is associated with an AI/ML model training. [0185] In one or more implementations, the UE feeds back CSI corresponding to the training dataset, where an indicator of the subset of the training dataset to which the fed back CSI belongs is included as part of the CSI report. In one example, a field corresponding to the indicator is reported in the CSI report based on a configured report quantity of the CSI Reporting setting. [0186] In another example, the indicator is implicit, e.g., not reported as a standalone parameter, and is instead inferred from an output value of a function that depends on coefficient values, spatial/frequency/time domain selected basis indices, e.g., based on an average gain corresponding to a group of coefficients corresponding to a subset of the dimensions. [0187] In another example, at least one value of a set of possible indicator values may correspond to a case where the CSI cannot be classified to any subset of the multiple subsets of the training dataset. E.g., there may be an indication that none of the AI/ML models work efficiently for a channel. [0188] In another example, the indicator value is based on multiple CSI-RS measurements, e.g., a sequence of CSI-RSs received over a periodic CSI-RS resource, or a semi-persistent CSI-RS resource configuration. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 54 [0189] In one or more implementations, model training is performed at the network side and shared with the UE. The network signals a set of parameters corresponding to multiple AI/ML models, e.g., encoder functions of a CSI auto-encoder, where the set of parameters are partitioned into multiple subsets of parameters, and each subset of parameters is associated with an AI/ML model of the multiple AI/ML models. In one example, the network signals the set of parameters as part of higher-layer, e.g., RRC signaling. [0190] In another example, the network signals the set of parameters as part of an AI-based report over at least one of PUSCH, PUCCH. [0191] In another example, the network signals a first subset of the set of parameters corresponding to a subset of the multiple AI/ML models in a first time unit and feeds back a second subset of the set of parameters corresponding to a subset of the multiple AI/ML models in a subsequent time unit. [0192] In another example, a subset of the set of parameters corresponding to a subset of the multiple AI/ML models is common for multiple network nodes. [0193] In another example, a subset of the set of parameters corresponding to a subset of the multiple AI/ML models is signaled to the UE based on an indicator value reported by the UE that indicates a characteristic that corresponds to the subset of the multiple AI/ML models. [0194] In one or more implementations, AI/ML model training is performed at the UE side and shared with the network. The UE signals a set of parameters corresponding to multiple AI/ML models, e.g., decoder functions of a CSI auto-encoder, where the set of parameters are partitioned into multiple subsets of parameters, each subset of parameters is associated with an AI/ML model of the multiple AI/ML models. In one example, the UE feeds back the set of parameters as part of a CSI report over at least one of PUSCH, PUCCH. [0195] In another example, the UE feeds back the set of parameters as part of an AI-based report over at least one of PUSCH, PUCCH. [0196] In another example, the UE feeds back a first subset of the set of parameters corresponding to a subset of the multiple AI/ML models in a first report and feeds back a second Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 55 subset of the set of parameters corresponding subset of the multiple AI/ML models in a second subsequent report. [0197] In another example, the UE feeds back the set of parameters as part of higher-layer signaling. [0198] In another example, a group of subsets of parameters is configured, and the UE provides the network with a set of parameters corresponding to each subset of the group of subsets of parameters. [0199] In one or more implementations, the UE receives a configuration signal, e.g., as part of higher-layer configuration signaling, that indicates a CSI-based metric that the UE would compute to select an AI/ML model from the multiple AI/ML models. In one example, the CSI-based metric corresponds to a number of frequency-domain basis indices that is reported by the UE. [0200] In another example, the CSI-based metric corresponds to a number of spatial-domain basis indices that is reported by the UE. [0201] In another example, the CSI-based metric corresponds to a ratio of a function of power (or alternatively amplitude) gain of a first subset of spatial-domain basis indices to a function of power (or alternatively amplitude) gain of a second subset of spatial-domain basis indices. [0202] In another example, the CSI-based metric corresponds to a ratio of a function of power (or alternatively amplitude) gain of a first subset of frequency-domain basis indices to a function of power (or alternatively amplitude) gain of a second subset of frequency-domain basis indices. [0203] In one or more implementations, the network indicates to the UE an index of an AI/ML model from the multiple AI/ML models based on uplink-downlink channel reciprocity. In one example, the network indicates signals an index of s selected AI/ML model as part of at least one of DCI signaling, MAC CE signaling, or RRC signaling. In another example, the uplink-downlink channel reciprocity corresponds to a sounding reference signal that is coupled with a CSI-RS resource via sounding reference signal (SRS) resource configuration signaling as part of a higher- layer signaling. [0204] In one or more implementations, the network indicates via a first indication the model parameters based on the training. The model parameters are applicable after a first certain time from Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 56 a time reference associated with the first The first certain time can be pre-determined, reported via a UE capability signaling, depends on a characteristic of training data set (e.g., related to a number of dominant channel paths). The UE does not expect to receive a second indication indicating (a second) model parameters prior to elapsing the first certain time. [0205] CSI feedback adjustment based on a selected AI/ML model is discussed below. As discussed above, the CSI feedback may be mismatched compared with the possible channel variations/distributions represented in the multiple AI/ML models, e.g., due to bursty interference, instantaneous hardware issues or instantaneous blockage. In light of that, some adjustment to the inferred output of the AI/ML model may be made to adjust with this instantaneous channel variation. Several implementations are described below. One or more elements or features from one or more of the described implementations may be combined. [0206] In one or more implementations, the UE can switch off some beams based on some instantaneous channel variation that is not captured in the model. The UE feeds back a bitmap corresponding to at least one of channel/precoder spatial domain dimensions, frequency domain dimensions and time domain dimensions, where the bitmap indicates whether at least one of the aforementioned dimensions is turned off, i.e., coefficients corresponding to the aforementioned dimensions are given a zero amplitude value, even if the inferred value corresponding to the AI/ML model is non-zero. [0207] In one or more implementations , UE can attenuate some beams based on some instantaneous channel variation that is not captured in the model. The UE feeds back a bitmap corresponding to at least one of channel/precoder spatial domain dimensions, frequency domain dimensions and time domain dimensions, where the bitmap indicates whether at least one of the aforementioned dimensions is scaled, e.g., coefficients corresponding to the aforementioned dimensions are given a scaled amplitude value, where the scaling is applied to the inferred values from the AI/ML model [0208] In one or more implementations, UE can attenuate/switch off beams based on code book subset restriction (CBSR). The adjustment of amplitude values of the subset of coefficients is in a form of a UE feedback corresponding to codebook subset restriction feedback. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 57 [0209] In one or more implementations, of amplitude values is network-based, e.g., the network signals the adjustment. [0210] In one or more implementations, the terms antenna, panel, and antenna panel are used interchangeably. An antenna panel may be a hardware that is used for transmitting and/or receiving radio signals at frequencies lower than 6GHz, e.g., frequency range 1 (FR1), or higher than 6GHz, e.g., frequency range 2 (FR2) or millimeter wave (mmWave). In some implementations, an antenna panel may comprise an array of antenna elements, where each antenna element is connected to hardware such as a phase shifter that allows a control module to apply spatial parameters for transmission and/or reception of signals. The resulting radiation pattern may be called a beam, which may or may not be unimodal and may allow the device to amplify signals that are transmitted or received from spatial directions. [0211] In one or more implementations, an antenna panel may or may not be virtualized as an antenna port. An antenna panel may be connected to a baseband processing module through a radio frequency (RF) chain for each of transmission (egress) and reception (ingress) directions. A capability of a device in terms of the number of antenna panels, their duplexing capabilities, their beamforming capabilities, and so on, may or may not be transparent to other devices. Capability information may be communicated via signaling or capability information may be provided to devices without a need for signaling. In the case that such information is available to other devices, it can be used for signaling or local decision making. [0212] In one or more implementations, a device (e.g., UE, node) antenna panel may be a physical or logical antenna array comprising a set of antenna elements or antenna ports that share a common or a significant portion of an RF chain (e.g., in-phase/quadrature (I/Q) modulator, analog to digital (A/D) converter, local oscillator, phase shift network). The device antenna panel or “device panel” may be a logical entity with physical device antennas mapped to the logical entity. The mapping of physical device antennas to the logical entity may be up to device implementation. Communicating (receiving or transmitting) on at least a subset of antenna elements or antenna ports active for radiating energy (also referred to herein as active elements) of an antenna panel requires biasing or powering on of the RF chain which results in current drain or power consumption in the device associated with the antenna panel (including power amplifier/low noise amplifier (LNA) power consumption associated with the antenna elements or antenna ports). The phrase "active for Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 58 radiating energy," as used herein, is not meant limited to a transmit function but also encompasses a receive function. Accordingly, an antenna element that is active for radiating energy may be coupled to a transmitter to transmit radio frequency energy or to a receiver to receive radio frequency energy, either simultaneously or sequentially, or may be coupled to a transceiver in general, for performing its intended functionality. Communicating on the active elements of an antenna panel enables generation of radiation patterns or beams. [0213] In one or more implementations, depending on device’s own implementation, a “device panel” can have at least one of the following functionalities as an operational role of Unit of antenna group to control its Tx beam independently, Unit of antenna group to control its transmission power independently, Unit of antenna group to control its transmission timing independently. The “device panel” may be transparent to the network entity (e.g., gNB). For certain condition(s), the network entity (e.g., gNB) or network can assume the mapping between device’s physical antennas to the logical entity “device panel” may not be changed. For example, the condition may include until the next update or report from device or comprise a duration of time over which the network entity (e.g., gNB) assumes there will be no change to the mapping. A device may report its capability with respect to the “device panel” to the network entity (e.g., gNB) or network. The device capability may include at least the number of “device panels”. In one or more implementations, the device may support uplink (UL) transmission from one beam within a panel; with multiple panels, more than one beam (one beam per panel) may be used for UL transmission. Additionally or alternatively, more than one beam per panel may be supported/used for UL transmission. [0214] In one or more implementations, an antenna port is defined such that the channel over which a symbol on the antenna port is conveyed can be inferred from the channel over which another symbol on the same antenna port is conveyed. [0215] Two antenna ports are said to be quasi co-located (QCL) if the large-scale properties of the channel over which a symbol on one antenna port is conveyed can be inferred from the channel over which a symbol on the other antenna port is conveyed. The large-scale properties include one or more of delay spread, Doppler spread, Doppler shift, average gain, average delay, and spatial Rx parameters. Two antenna ports may be quasi-located with respect to a subset of the large-scale properties and different subset of large-scale properties may be indicated by a QCL Type. The QCL Type can indicate which channel properties are the same between the two reference signals (e.g., on Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 59 the two antenna ports). Thus, the reference can be linked to each other with respect to what the UE can assume about their channel statistics or QCL properties. For example, qcl-Type may take one of the following values: 'QCL-TypeA': {Doppler shift, Doppler spread, average delay, delay spread}; 'QCL-TypeB': {Doppler shift, Doppler spread}; 'QCL-TypeC': {Doppler shift, average delay}; 'QCL-TypeD': {Spatial Rx parameter}. [0216] Spatial Rx parameters may include one or more of: angle of arrival (AoA,) Dominant AoA, average AoA, angular spread, Power Angular Spectrum (PAS) of AoA, average AoD (angle of departure), PAS of AoD, transmit/receive channel correlation, transmit/receive beamforming, spatial channel correlation etc. [0217] The QCL-TypeA, QCL-TypeB and QCL-TypeC may be applicable for all carrier frequencies, but the QCL-TypeD may be applicable only in higher carrier frequencies (e.g., mmWave, FR2 and beyond), where essentially the UE may not be able to perform omni-directional transmission, e.g., the UE would need to form beams for directional transmission. A QCL-TypeD between two reference signals A and B, the reference signal A is considered to be spatially co- located with reference signal B and the UE may assume that the reference signals A and B can be received with the same spatial filter (e.g., with the same receive (RX) beamforming weights). [0218] An “antenna port” according to one or more implementations may be a logical port that may correspond to a beam (resulting from beamforming) or may correspond to a physical antenna on a device. In one or more implementations, a physical antenna may map directly to a single antenna port, in which an antenna port corresponds to an actual physical antenna. Additionally or alternatively, a set or subset of physical antennas, or antenna set or antenna array or antenna sub- array, may be mapped to one or more antenna ports after applying complex weights, a cyclic delay, or both to the signal on each physical antenna. The physical antenna set may have antennas from a single module or panel or from multiple modules or panels. The weights may be fixed as in an antenna virtualization scheme, such as cyclic delay diversity (CDD). The procedure used to derive antenna ports from physical antennas may be specific to a device implementation and transparent to other devices. [0219] In one or more implementations, a TCI-state (Transmission Configuration Indication) associated with a target transmission can indicate parameters for configuring a quasi-collocation Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 60 relationship between the target transmission target RS of demodulation reference signal (DM- RS) ports of the target transmission during a transmission occasion) and a source reference signal(s) (e.g., SSB/CSI-RS/SRS) with respect to quasi co-location type parameter(s) indicated in the corresponding TCI state. The TCI describes which reference signals are used as QCL source, and what QCL properties can be derived from each reference signal. A device can receive a configuration of multiple transmission configuration indicator states for a serving cell for transmissions on the serving cell. In one or more implementations, a TCI state comprises at least one source RS to provide a reference (UE assumption) for determining QCL and/or spatial filter. [0220] In one or more implementations, a spatial relation information associated with a target transmission can indicate parameters for configuring a spatial setting between the target transmission and a reference RS (e.g., SSB/CSI-RS/SRS). For example, the device may transmit the target transmission with the same spatial domain filter used for reception the reference RS (e.g., DL RS such as SSB/CSI-RS). In another example, the device may transmit the target transmission with the same spatial domain transmission filter used for the transmission of the reference RS (e.g., UL RS such as SRS). A device can receive a configuration of multiple spatial relation information configurations for a serving cell for transmissions on the serving cell. [0221] In one or more implementations, a UL TCI state is provided if a device is configured with separate DL/UL TCI by RRC signaling. The UL TCI state may comprise a source reference signal which provides a reference for determining UL spatial domain transmission filter for the UL transmission (e.g., dynamic-grant/configured-grant based PUSCH, dedicated PUCCH resources) in a component carrier (CC) or across a set of configured CCs/bandwidth parts (BWPs). [0222] In one or more implementations, a joint DL/UL TCI state is provided if the device is configured with joint DL/UL TCI by RRC signaling (e.g., configuration of joint TCI or separate DL/UL TCI is based on RRC signaling). The joint DL/UL TCI state refers to at least a common source reference RS used for determining both the DL QCL information and the UL spatial transmission filter. The source RS determined from the indicated joint (or common) TCI state provides QCL Type-D indication (e.g., for device-dedicated physical downlink control channel (PDCCH)/physical downlink shared channel (PDSCH) and is used to determine UL spatial transmission filter (e.g., for UE-dedicated PUSCH/PUCCH) for a CC or across a set of configured CCs/BWPs. In one example, the UL spatial transmission filter is derived from the RS of DL QCL Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 61 Type D in the joint TCI state. The spatial the UL transmission may be according to the spatial relation with a reference to the source RS configured with qcl-Type set to 'typeD' in the joint TCI state. [0223] Accordingly, the techniques discussed herein provide an AI-based CSI feedback mechanism that provides a channel-matching precoder under different channel conditions, e.g., LoS/NLoS, Outdoor/Indoor status of a UE, and so forth. In order to accommodate for such variations in channel conditions, some flexibility in the precoder design is used. More specifically, the UE is configured with multiple AI/ML models, where each AI/ML model of the multiple AI/ML models is based on a distinct data set corresponding to a distribution of the CSI, such that the UE can toggle between different precoder selections corresponding to different channel distributions based on variations in UE location, orientation, outdoor/indoor, LoS/NLoS status, and so forth. Signaling between the UE and the network that indicates a selected AI/ML model from the plurality of AI/ML models based on a channel-based threshold is also discussed. [0224] Additionally or alternatively, the UE is configured with reporting an indication that corresponds to switching off a subset of the set of ports, a subset of the set of frequency sub-bands, or a combination thereof, such that the precoder can be adjusted based on instantaneous deviation of the channel from its typical distribution, e.g., due to bursty interference or instantaneous blockage. [0225] FIG. 8 illustrates an example of a block diagram 800 of a device 802 (e.g., an apparatus) that supports AI for CSI in accordance with aspects of the present disclosure. The device 802 may be an example of UE 104 as described herein. The device 802 may support wireless communication with one or more network entities 102, UE 104, or any combination thereof. The device 802 may include components for bi-directional communications including components for transmitting and receiving communications, such as a processor 804, a memory 806, a transceiver 808, and an I/O controller 810. These components may be in electronic communication or otherwise coupled (e.g., operatively, communicatively, functionally, electronically, electrically) via one or more interfaces (e.g., buses). [0226] The processor 804, the memory 806, the transceiver 808, or various combinations thereof or various components thereof may be examples of means for performing various aspects of the present disclosure as described herein. For example, the processor 804, the memory 806, the Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 62 transceiver 808, or various combinations or thereof may support a method for performing one or more of the operations described herein. [0227] In some implementations, the processor 804, the memory 806, the transceiver 808, or various combinations or components thereof may be implemented in hardware (e.g., in communications management circuitry). The hardware may include a processor, a digital signal processor (DSP), an application-specific integrated circuit (ASIC), a field-programmable gate array (FPGA) or other programmable logic device, a discrete gate or transistor logic, discrete hardware components, or any combination thereof configured as or otherwise supporting a means for performing the functions described in the present disclosure. In some implementations, the processor 804 and the memory 806 coupled with the processor 804 may be configured to perform one or more of the functions described herein (e.g., executing, by the processor 804, instructions stored in the memory 806). In the context of UE 104, for example, the transceiver 808 and the processor coupled 804 coupled to the transceiver 808 are configured to cause the UE 104 to perform the various described operations and/or combinations thereof. [0228] For example, the processor 804 and/or the transceiver 808 may support wireless communication at the device 802 in accordance with examples as disclosed herein. For instance, the processor 804 and/or the transceiver 808 may be configured as and/or otherwise support a means to generate at least one latent representation of input data based on at least one set of neural network models; generate at least one quantized representation of the at least one latent representation based on at least one of scalar quantization or vector quantization associated with the at least one set of neural network models; and transmit the at least one quantized representation. [0229] Further, in some implementations, the processor and the transceiver are configured to cause the apparatus (e.g., the device 802) to determine at least one of a quantization codebook corresponding to the vector quantization, a type of the scalar quantization, or a number of quantization levels for the scalar quantization; the processor and the transceiver are configured to cause the apparatus to receive an indication from a different apparatus, the indication including the at least one of the quantization codebook corresponding to the vector quantization, the type of the scalar quantization, or the number of quantization levels for the scalar quantization; the processor and the transceiver are configured to cause the apparatus to: determine a first latent representation of the at least one latent representation based on a first set of neural network models; determine a Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 63 second latent representation of the at least one representation based on a second set of neural network models; and a first quantized representation of the at least one quantized representation is based on vector quantization of the first latent representation, and a second quantized representation of the at least one quantized representation is based on scalar quantization of the second latent representation; and transmit the first quantized representation and the second quantized representation. [0230] Further, in some implementations, the first set of neural network models and the second set of neural network models include at least one common neural network model; the processor and the transceiver are configured to cause the apparatus to determine the at least one set of neural network models from a plurality of set of neural network models based on an indication from a second apparatus; the processor and the transceiver are configured to cause the apparatus to determine model configuration information for the at least one set of neural network models based on an indication from a second apparatus; the model configuration information includes at least one of a structure of at least one neural network of the at least one set of neural network models, or weights of at least one neural network of the at least one set of neural network models; the processor and the transceiver are configured to cause the apparatus to: select the at least one set of neural network models from a plurality of sets of neural network models; and transmit an indication of the selected at least one set of neural network models to a second apparatus. [0231] Further, in some implementations, the input data is based at least in part on a channel data representation; the processor and the transceiver are configured to cause the apparatus to determine the channel data representation based at least in part on at least one reference signal from a second apparatus; the processor and the transceiver are configured to cause the apparatus to determine the channel data representation based at least in part on at least one of different transmitter and receiver pairs over different frequency bands, different time slots, or time slot transformation in different domains; the at least one set of neural network models includes: a first neural network model corresponding to a first configuration of a parameter; and a second neural network model corresponding to a second configuration of the parameter; the parameter includes at least one of a codebook parameter specifying permitted codewords or a rank parameter specifying permitted ranks associated with the at least one quantized representation; the processor and the Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 64 transceiver are configured to cause the apply the first neural network model after a time duration from the first configuration of the parameter. [0232] Further, at least some implementations include a processor for wireless communication, including at least one controller coupled with at least one memory and configured to cause the processor to generate at least one latent representation of input data based on at least one set of neural network models, generate at least one quantized representation of the at least one latent representation based on at least one of scalar quantization or vector quantization associated with the at least one set of neural network models, and transmit the at least one quantized representation. [0233] The processor 804 may include an intelligent hardware device (e.g., a general-purpose processor, a DSP, a CPU, a microcontroller, an ASIC, an FPGA, a programmable logic device, a discrete gate or transistor logic component, a discrete hardware component, or any combination thereof). In some implementations, the processor 804 may be configured to operate a memory array using a memory controller. In some other implementations, a memory controller may be integrated into the processor 804. The processor 804 may be configured to execute computer-readable instructions stored in a memory (e.g., the memory 806) to cause the device 802 to perform various functions of the present disclosure. [0234] The memory 806 may include random access memory (RAM) and read-only memory (ROM). The memory 806 may store computer-readable, computer-executable code including instructions that, when executed by the processor 804 cause the device 802 to perform various functions described herein. The code may be stored in a non-transitory computer-readable medium such as system memory or another type of memory. In some implementations, the code may not be directly executable by the processor 804 but may cause a computer (e.g., when compiled and executed) to perform functions described herein. In some implementations, the memory 806 may include, among other things, a basic I/O system (BIOS) which may control basic hardware or software operation such as the interaction with peripheral components or devices. [0235] The I/O controller 810 may manage input and output signals for the device 802. The I/O controller 810 may also manage peripherals not integrated into the device M02. In some implementations, the I/O controller 810 may represent a physical connection or port to an external peripheral. In some implementations, the I/O controller 810 may utilize an operating system such as Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 65 iOS®, ANDROID®, MS-DOS®, MS- OS/2®, UNIX®, LINUX®, or another known operating system. In some implementations, the I/O controller 810 may be implemented as part of a processor, such as the processor M08. In some implementations, a user may interact with the device 802 via the I/O controller 810 or via hardware components controlled by the I/O controller 810. [0236] In some implementations, the device 802 may include a single antenna 812. However, in some other implementations, the device 802 may have more than one antenna 812 (e.g., multiple antennas), including multiple antenna panels or antenna arrays, which may be capable of concurrently transmitting or receiving multiple wireless transmissions. The transceiver 808 may communicate bi-directionally, via the one or more antennas 812, wired, or wireless links as described herein. For example, the transceiver 808 may represent a wireless transceiver and may communicate bi-directionally with another wireless transceiver. The transceiver 808 may also include a modem to modulate the packets, to provide the modulated packets to one or more antennas 812 for transmission, and to demodulate packets received from the one or more antennas 812. [0237] FIG. 9 illustrates an example of a block diagram 900 of a device 902 (e.g., an apparatus) that supports AI for CSI in accordance with aspects of the present disclosure. The device 902 may be an example of a network entity 102 as described herein. The device 902 may support wireless communication with one or more network entities 102, UE 104, or any combination thereof. The device 902 may include components for bi-directional communications including components for transmitting and receiving communications, such as a processor 904, a memory 906, a transceiver 908, and an I/O controller 910. These components may be in electronic communication or otherwise coupled (e.g., operatively, communicatively, functionally, electronically, electrically) via one or more interfaces (e.g., buses). [0238] The processor 904, the memory 906, the transceiver 908, or various combinations thereof or various components thereof may be examples of means for performing various aspects of the present disclosure as described herein. For example, the processor 904, the memory 906, the transceiver 908, or various combinations or components thereof may support a method for performing one or more of the operations described herein. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 66 [0239] In some implementations, the 904, the memory 906, the transceiver 908, or various combinations or components thereof may be implemented in hardware (e.g., in communications management circuitry). The hardware may include a processor, a digital signal processor (DSP), an application-specific integrated circuit (ASIC), a field-programmable gate array (FPGA) or other programmable logic device, a discrete gate or transistor logic, discrete hardware components, or any combination thereof configured as or otherwise supporting a means for performing the functions described in the present disclosure. In some implementations, the processor 904 and the memory 906 coupled with the processor 904 may be configured to perform one or more of the functions described herein (e.g., executing, by the processor 904, instructions stored in the memory 906). In the context of network entity 102, for example, the transceiver 908 and the processor 904 coupled to the transceiver 908 are configured to cause the network entity 102 to perform the various described operations and/or combinations thereof. [0240] For example, the processor 904 and/or the transceiver 908 may support wireless communication at the device 902 in accordance with examples as disclosed herein. For instance, the processor 904 and/or the transceiver 908 may be configured as or otherwise support a means to receive at least one set of data; determine at least one latent representation of the at least one set of data; and determine an output using the at least one latent representation. [0241] Further, in some implementations, to determine the at least one latent representation, the processor and the transceiver are configured to cause the apparatus (e.g., the device 902) to determine at least one set of neural network models from a plurality of sets of neural network models and based on an indication from a second apparatus; the processor and the transceiver are configured to cause the apparatus to determine model configuration information of the at least one set of neural network models based on an indication received from a second apparatus; the model configuration information includes at least one of a structure of at least one neural network of the at least one set of neural network models, or weights of at least one neural network of the at least one set of neural network models; the at least one set of data includes a first set of data and a second set of data, and the processor and the transceiver are configured to cause the apparatus to: determine a first latent representation of the at least one latent representation based on at least a quantization codebook and the first set of data; determine a second latent representation of the at least one latent representation based on the second set of data; and determine the output using the first latent Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 67 representation and the second latent the processor and the transceiver are configured to cause the apparatus to determine the quantization codebook as a quantization codebook corresponding to a demapper module; the processor and the transceiver are configured to cause the apparatus to determine the quantization codebook based on an indication received from a different apparatus. [0242] The processor 904 may include an intelligent hardware device (e.g., a general-purpose processor, a DSP, a CPU, a microcontroller, an ASIC, an FPGA, a programmable logic device, a discrete gate or transistor logic component, a discrete hardware component, or any combination thereof). In some implementations, the processor 904 may be configured to operate a memory array using a memory controller. In some other implementations, a memory controller may be integrated into the processor 904. The processor 904 may be configured to execute computer-readable instructions stored in a memory (e.g., the memory 906) to cause the device 902 to perform various functions of the present disclosure. [0243] The memory 906 may include random access memory (RAM) and read-only memory (ROM). The memory 906 may store computer-readable, computer-executable code including instructions that, when executed by the processor 904 cause the device 902 to perform various functions described herein. The code may be stored in a non-transitory computer-readable medium such as system memory or another type of memory. In some implementations, the code may not be directly executable by the processor 904 but may cause a computer (e.g., when compiled and executed) to perform functions described herein. In some implementations, the memory 906 may include, among other things, a basic I/O system (BIOS) which may control basic hardware or software operation such as the interaction with peripheral components or devices. [0244] The I/O controller 910 may manage input and output signals for the device 902. The I/O controller 910 may also manage peripherals not integrated into the device M02. In some implementations, the I/O controller 910 may represent a physical connection or port to an external peripheral. In some implementations, the I/O controller 910 may utilize an operating system such as iOS®, ANDROID®, MS-DOS®, MS-WINDOWS®, OS/2®, UNIX®, LINUX®, or another known operating system. In some implementations, the I/O controller 910 may be implemented as part of a processor, such as the processor M06. In some implementations, a user may interact with Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 68 the device 902 via the I/O controller 910 or via components controlled by the I/O controller 910. [0245] In some implementations, the device 902 may include a single antenna 912. However, in some other implementations, the device 902 may have more than one antenna 912 (e.g., multiple antennas), including multiple antenna panels or antenna arrays, which may be capable of concurrently transmitting or receiving multiple wireless transmissions. The transceiver 908 may communicate bi-directionally, via the one or more antennas 912, wired, or wireless links as described herein. For example, the transceiver 908 may represent a wireless transceiver and may communicate bi-directionally with another wireless transceiver. The transceiver 908 may also include a modem to modulate the packets, to provide the modulated packets to one or more antennas 912 for transmission, and to demodulate packets received from the one or more antennas 912. [0246] FIG. 10 illustrates a flowchart of a method 1000 that supports AI for CSI in accordance with aspects of the present disclosure. The operations of the method 1000 may be implemented by a device or its components as described herein. For example, the operations of the method 1000 may be performed by a UE 104 as described with reference to FIGs. 1 through 9. In some implementations, the device may execute a set of instructions to control the function elements of the device to perform the described functions. Additionally, or alternatively, the device may perform aspects of the described functions using special-purpose hardware. [0247] At 1002, the method may include generating at a first apparatus at least one latent representation of input data based on at least one set of neural network models. The operations of 1002 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1002 may be performed by a device as described with reference to FIG. 1. [0248] At 1004, the method may include generating at least one quantized representation of the at least one latent representation based on at least one of scalar quantization or vector quantization associated with the at least one set of neural network models. The operations of 1004 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1004 may be performed by a device as described with reference to FIG. 1. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 69 [0249] At 1006, the method may include the at least one quantized representation. The operations of 1006 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1006 may be performed by a device as described with reference to FIG. 1. [0250] FIG. 11 illustrates a flowchart of a method 1100 that supports AI for CSI in accordance with aspects of the present disclosure. The operations of the method 1100 may be implemented by a device or its components as described herein. For example, the operations of the method 1100 may be performed by a UE 104 as described with reference to FIGs. 1 through 9. In some implementations, the device may execute a set of instructions to control the function elements of the device to perform the described functions. Additionally, or alternatively, the device may perform aspects of the described functions using special-purpose hardware. [0251] At 1102, the method may include determining a first latent representation of the at least one latent representation based on a first set of neural network models. The operations of 1102 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1102 may be performed by a device as described with reference to FIG. 1. [0252] At 1104, the method may include determining a second latent representation of the at least one latent representation based on a second set of neural network models; and where: a first quantized representation of the at least one quantized representation is based on vector quantization of the first latent representation, and a second quantized representation of the at least one quantized representation is based on scalar quantization of the second latent representation. The operations of 1104 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1104 may be performed by a device as described with reference to FIG. 1. [0253] At 1106, the method may include transmitting the first quantized representation and the second quantized representation. The operations of 1106 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1106 may be performed by a device as described with reference to FIG. 1. [0254] FIG. 12 illustrates a flowchart of a method 1200 that supports AI for CSI in accordance with aspects of the present disclosure. The operations of the method 1200 may be implemented by a Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 70 device or its components as described herein. example, the operations of the method 1200 may be performed by a UE 104 as described with reference to FIGs. 1 through 9. In some implementations, the device may execute a set of instructions to control the function elements of the device to perform the described functions. Additionally, or alternatively, the device may perform aspects of the described functions using special-purpose hardware. [0255] At 1202, the method may include selecting the at least one set of neural network models from a plurality of sets of neural network models. The operations of 1202 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1202 may be performed by a device as described with reference to FIG. 1. [0256] At 1204, the method may include transmitting an indication of the selected at least one set of neural network models to a second apparatus. The operations of 1204 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1204 may be performed by a device as described with reference to FIG. 1. [0257] FIG. 13 illustrates a flowchart of a method 1300 that supports AI for CSI in accordance with aspects of the present disclosure. The operations of the method 1300 may be implemented by a device or its components as described herein. For example, the operations of the method 1300 may be performed by a network entity 102 as described with reference to FIGs. 1 through 9. In some implementations, the device may execute a set of instructions to control the function elements of the device to perform the described functions. Additionally, or alternatively, the device may perform aspects of the described functions using special-purpose hardware. [0258] At 1302, the method may include receiving at a first apparatus at least one set of data. The operations of 1302 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1302 may be performed by a device as described with reference to FIG. 1. [0259] At 1304, the method may include determining at least one latent representation of the at least one set of data. The operations of 1304 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1304 may be performed by a device as described with reference to FIG. 1. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 71 [0260] At 1306, the method may include an output using the at least one latent representation. The operations of 1306 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1306 may be performed by a device as described with reference to FIG. 1. [0261] FIG. 14 illustrates a flowchart of a method 1400 that supports AI for CSI in accordance with aspects of the present disclosure. The operations of the method 1400 may be implemented by a device or its components as described herein. For example, the operations of the method 1400 may be performed by a network entity 102 as described with reference to FIGs. 1 through 9. In some implementations, the device may execute a set of instructions to control the function elements of the device to perform the described functions. Additionally, or alternatively, the device may perform aspects of the described functions using special-purpose hardware. [0262] At 1402, the method may include determining a first latent representation of the at least one latent representation based on at least a quantization codebook and the first set of data. The operations of 1402 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1402 may be performed by a device as described with reference to FIG. 1. [0263] At 1404, the method may include determining a second latent representation of the at least one latent representation based on the second set of data. The operations of 1404 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1404 may be performed by a device as described with reference to FIG. 1. [0264] At 1406, the method may include determining the output using the first latent representation and the second latent representation. The operations of 1406 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1406 may be performed by a device as described with reference to FIG. 1. [0265] FIG. 15 illustrates a flowchart of a method 1500 that supports AI for CSI in accordance with aspects of the present disclosure. The operations of the method 1500 may be implemented by a device or its components as described herein. For example, the operations of the method 1500 may be performed by a network entity 102 as described with reference to FIGs. 1 through 9. In some implementations, the device may execute a set of instructions to control the function elements of the Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 72 device to perform the described functions. or alternatively, the device may perform aspects of the described functions using special-purpose hardware. [0266] At 1502, the method may include receiving at a first apparatus a first data set from a second apparatus. The operations of 1502 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1502 may be performed by a device as described with reference to FIG. 1. [0267] At 1504, the method may include selecting, from a plurality of two-sided models and based at least in part on the first data set, a two-sided model comprising an encoder model and a decoder model. The operations of 1504 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1504 may be performed by a device as described with reference to FIG. 1. [0268] At 1506, the method may include transmitting, to the second apparatus, at least one encoder parameter for the encoder model. The operations of 1506 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1506 may be performed by a device as described with reference to FIG. 1. [0269] At 1508, the method may include receiving, from the second apparatus, feedback data based at least in part on encoder model. The operations of 1508 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1508 may be performed by a device as described with reference to FIG. 1. [0270] At 1510, the method may include generating output data based at least in part on the decoder model and at least a portion of the feedback data. The operations of 1510 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1510 may be performed by a device as described with reference to FIG. 1. [0271] FIG. 16 illustrates a flowchart of a method 1600 that supports AI for CSI in accordance with aspects of the present disclosure. The operations of the method 1600 may be implemented by a device or its components as described herein. For example, the operations of the method 1600 may be performed by a UE 104 as described with reference to FIGs. 1 through 9. In some implementations, the device may execute a set of instructions to control the function elements of the Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 73 device to perform the described functions. or alternatively, the device may perform aspects of the described functions using special-purpose hardware. [0272] At 1602, the method may include receiving, at a first apparatus and from a second apparatus, at least one configuration parameter for a two-sided model including at least one encoder parameter for an encoder of the two-sided model, wherein the two-sided model comprises at least one set of neural network models. The operations of 1602 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1602 may be performed by a device as described with reference to FIG. 1. [0273] At 1604, the method may include generating a latent representation based at least in part on input data and a version of the at least one encoder parameter. The operations of 1604 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1604 may be performed by a device as described with reference to FIG. 1. [0274] At 1606, the method may include generating feedback data comprising a quantization of the latent representation based on a quantization scheme. The operations of 1606 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1606 may be performed by a device as described with reference to FIG. 1. [0275] At 1608, the method may include transmitting, to the second apparatus, the feedback data. The operations of 1608 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1608 may be performed by a device as described with reference to FIG. 1. [0276] FIG. 17 illustrates a flowchart of a method 1700 that supports AI for CSI in accordance with aspects of the present disclosure. The operations of the method 1700 may be implemented by a device or its components as described herein. For example, the operations of the method 1700 may be performed by a UE 104 such as described with reference to FIGs. 1 through 9. In some implementations, the device may execute a set of instructions to control the function elements of the device to perform the described functions. Additionally, or alternatively, the device may perform aspects of the described functions using special-purpose hardware. [0277] At 1702, the method may include selecting the two-sided model from the plurality of two-sided models and based at least in part on the selection neural network and the input data. The Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 74 operations of 1702 may be performed in with examples as described herein. In some implementations, aspects of the operations of 1702 may be performed by a device as described with reference to FIG. 1. [0278] At 1704, the method may include transmitting an indication of the two-sided model to the second apparatus. The operations of 1704 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1704 may be performed by a device as described with reference to FIG. 1. [0279] FIG. 18 illustrates a flowchart of a method 1800 that supports generating a measurement report using one of multiple available artificial intelligence models in accordance with aspects of the present disclosure. The operations of the method 1800 may be implemented by a device or its components as described herein. For example, the operations of the method 1800 may be performed by a UE 104 as described with reference to FIGs. 1-9. In some implementations, the device may execute a set of instructions to control the function elements of the device to perform the described functions. Additionally, or alternatively, the device may perform aspects of the described functions using special-purpose hardware. [0280] At 1805, the method may include receiving, from a network entity, a first signaling indicating a configuration for measurement and reporting of at least one quantity, the configuration indicating a set of reference signals for measurement of the at least one quantity and indicating to report one or more parameters corresponding to the at least one quantity. The operations of 1805 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1805 may be performed by a device as described with reference to FIG. 1. [0281] At 1810, the method may include generating a measurement report, including the one or more parameters corresponding to the at least one quantity, based at least in part on the set of reference signals and a selection of an AI model from multiple AI models, each AI model of the multiple AI models having been configured prior to receiving the first signaling. The operations of 1810 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1810 may be performed by a device as described with reference to FIG. 1. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 75 [0282] At 1815, the method may include to the network entity, a second signaling indicating the measurement report. The operations of 1815 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1815 may be performed by a device as described with reference to FIG. 1. [0283] FIG. 19 illustrates a flowchart of a method 1900 that supports generating a measurement report using one of multiple available artificial intelligence models in accordance with aspects of the present disclosure. The operations of the method 1900 may be implemented by a device or its components as described herein. For example, the operations of the method 1900 may be performed by a UE 104 as described with reference to FIGs. 1-9. In some implementations, the device may execute a set of instructions to control the function elements of the device to perform the described functions. Additionally, or alternatively, the device may perform aspects of the described functions using special-purpose hardware. [0284] At 1905, the method may include receiving, from the network entity, a third signaling indicating one or more parameters corresponding to the selection of the AI model from the multiple AI models. The operations of 1905 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1905 may be performed by a device as described with reference to FIG. 1. [0285] At 1910, the method may include generating the measurement report based at least in part on the AI model and the one or more parameters that correspond to the AI model. The operations of 1910 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 1910 may be performed by a device as described with reference to FIG. 1. [0286] FIG. 20 illustrates a flowchart of a method 2000 that supports generating a measurement report using one of multiple available artificial intelligence models in accordance with aspects of the present disclosure. The operations of the method 2000 may be implemented by a device or its components as described herein. For example, the operations of the method 2000 may be performed by a network entity 102 as described with reference to FIGs. 1-9. In some implementations, the device may execute a set of instructions to control the function elements of the device to perform Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 76 the described functions. Additionally, or the device may perform aspects of the described functions using special-purpose hardware. [0287] At 2005, the method may include transmitting, to a UE, a first signaling indicating a configuration of the UE for measurement and reporting of at least one quantity, the configuration indicating a set of reference signals for measurement of the at least one quantity and indicating to report one or more parameters corresponding to the at least one quantity. The operations of 2005 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 2005 may be performed by a device as described with reference to FIG. 1. [0288] At 2010, the method may include receiving, from the UE, a second signaling indicating a measurement report, including the one or more parameters corresponding to the at least one quantity, generated based at least in part on the set of reference signals and a selection of an AI model from multiple AI models, each AI model of the multiple AI models having been configured prior to receiving the first signaling. The operations of 2010 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 2010 may be performed by a device as described with reference to FIG. 1. [0289] FIG. 21 illustrates a flowchart of a method 2100 that supports generating a measurement report using one of multiple available artificial intelligence models in accordance with aspects of the present disclosure. The operations of the method 2100 may be implemented by a device or its components as described herein. For example, the operations of the method 2100 may be performed by a network entity 102 as described with reference to FIGs. 1-9. In some implementations, the device may execute a set of instructions to control the function elements of the device to perform the described functions. Additionally, or alternatively, the device may perform aspects of the described functions using special-purpose hardware. [0290] At 2105, the method may include transmitting, to the UE, a third signaling indicating one or more parameters corresponding to the selection of the AI model from the multiple AI models. The operations of 2105 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 2105 may be performed by a device as described with reference to FIG. 1. Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 77 [0291] At 2110, the method may include from the UE, the measurement report generated based at least in part on the AI model and the one or more parameters that correspond to the one AI model. The operations of 2110 may be performed in accordance with examples as described herein. In some implementations, aspects of the operations of 2110 may be performed by a device as described with reference to FIG. 1. [0292] It should be noted that the methods described herein describes possible implementations, and that the operations and the steps may be rearranged or otherwise modified and that other implementations are possible. Further, aspects from two or more of the methods may be combined. [0293] The various illustrative blocks and components described in connection with the disclosure herein may be implemented or performed with a general-purpose processor, a DSP, an ASIC, a CPU, an FPGA or other programmable logic device, discrete gate or transistor logic, discrete hardware components, or any combination thereof designed to perform the functions described herein. A general-purpose processor may be a microprocessor, but in the alternative, the processor may be any processor, controller, microcontroller, or state machine. A processor may also be implemented as a combination of computing devices (e.g., a combination of a DSP and a microprocessor, multiple microprocessors, one or more microprocessors in conjunction with a DSP core, or any other such configuration. [0294] The functions described herein may be implemented in hardware, software executed by a processor, firmware, or any combination thereof. If implemented in software executed by a processor, the functions may be stored on or transmitted over as one or more instructions or code on a computer-readable medium. Other examples and implementations are within the scope of the disclosure and appended claims. For example, due to the nature of software, functions described herein may be implemented using software executed by a processor, hardware, firmware, hardwiring, or combinations of any of these. Features implementing functions may also be physically located at various positions, including being distributed such that portions of functions are implemented at different physical locations. [0295] Computer-readable media includes both non-transitory computer storage media and communication media including any medium that facilitates transfer of a computer program from one place to another. A non-transitory storage medium may be any available medium that may be Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 78 accessed by a general-purpose or special- computer. By way of example, and not limitation, non-transitory computer-readable media may include RAM, ROM, electrically erasable programmable ROM (EEPROM), flash memory, compact disk (CD) ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other non-transitory medium that may be used to carry or store desired program code means in the form of instructions or data structures and that may be accessed by a general-purpose or special-purpose computer, or a general-purpose or special-purpose processor. [0296] Any connection may be properly termed a computer-readable medium. For example, if the software is transmitted from a website, server, or other remote source using a coaxial cable, fiber optic cable, twisted pair, digital subscriber line (DSL), or wireless technologies such as infrared, radio, and microwave, then the coaxial cable, fiber optic cable, twisted pair, DSL, or wireless technologies such as infrared, radio, and microwave are included in the definition of computer-readable medium. Disk and disc, as used herein, include CD, laser disc, optical disc, digital versatile disc (DVD), floppy disk and Blu-ray disc where disks usually reproduce data magnetically, while discs reproduce data optically with lasers. Combinations of the above are also included within the scope of computer-readable media. [0297] As used herein, including in the claims, “or” as used in a list of items (e.g., a list of items prefaced by a phrase such as “at least one of” or “one or more of” or “one or both of”) indicates an inclusive list such that, for example, a list of at least one of A, B, or C means A or B or C or AB or AC or BC or ABC (e.g., A and B and C). Also, as used herein, the phrase “based on” shall not be construed as a reference to a closed set of conditions. For example, an example step that is described as “based on condition A” may be based on both a condition A and a condition B without departing from the scope of the present disclosure. In other words, as used herein, the phrase “based on” shall be construed in the same manner as the phrase “based at least in part on. Further, as used herein, including in the claims, a “set” may include one or more elements. [0298] The terms “transmitting,” “receiving,” or “communicating,” when referring to a network entity, may refer to any portion of a network entity (e.g., a base station, a CU, a DU, a RU) of a RAN communicating with another device (e.g., directly or via one or more other network entities). Attorney Docket No. SMM920220072-WO-PCT Lenovo Docket No. SMM920220072-WO-PCT 79 [0299] The description set forth herein, in with the appended drawings, describes example configurations and does not represent all the examples that may be implemented or that are within the scope of the claims. The term “example” used herein means “serving as an example, instance, or illustration,” and not “preferred” or “advantageous over other examples.” The detailed description includes specific details for the purpose of providing an understanding of the described techniques. These techniques, however, may be practiced without these specific details. In some instances, known structures and devices are shown in block diagram form to avoid obscuring the concepts of the described example. [0300] The description herein is provided to enable a person having ordinary skill in the art to make or use the disclosure. Various modifications to the disclosure will be apparent to a person having ordinary skill in the art, and the generic principles defined herein may be applied to other variations without departing from the scope of the disclosure. Thus, the disclosure is not limited to the examples and designs described herein but is to be accorded the broadest scope consistent with the principles and novel features disclosed herein. Attorney Docket No. SMM920220072-WO-PCT