1 /*
2 * Copyright 2018 Advanced Micro Devices, Inc.
3  *
4  * Permission is hereby granted, free of charge, to any person obtaining a
5  * copy of this software and associated documentation files (the "Software"),
6  * to deal in the Software without restriction, including without limitation
7  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8  * and/or sell copies of the Software, and to permit persons to whom the
9  * Software is furnished to do so, subject to the following conditions:
10  *
11  * The above copyright notice and this permission notice shall be included in
12  * all copies or substantial portions of the Software.
13  *
14  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
17  * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18  * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19  * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20  * OTHER DEALINGS IN THE SOFTWARE.
21  *
22  * Authors: AMD
23  *
24  */
25 #include <linux/delay.h>
26 #include "dm_services.h"
27 #include "dcn20/dcn20_hubbub.h"
28 #include "dcn21_hubbub.h"
29 #include "reg_helper.h"
30 
31 #define REG(reg)\
32 	hubbub1->regs->reg
33 #define DC_LOGGER \
34 	hubbub1->base.ctx->logger
35 #define CTX \
36 	hubbub1->base.ctx
37 
38 #undef FN
39 #define FN(reg_name, field_name) \
40 	hubbub1->shifts->field_name, hubbub1->masks->field_name
41 
42 #define REG(reg)\
43 	hubbub1->regs->reg
44 
45 #define CTX \
46 	hubbub1->base.ctx
47 
48 #undef FN
49 #define FN(reg_name, field_name) \
50 	hubbub1->shifts->field_name, hubbub1->masks->field_name
51 
convert_and_clamp(uint32_t wm_ns,uint32_t refclk_mhz,uint32_t clamp_value)52 static uint32_t convert_and_clamp(
53 	uint32_t wm_ns,
54 	uint32_t refclk_mhz,
55 	uint32_t clamp_value)
56 {
57 	uint32_t ret_val = 0;
58 	ret_val = wm_ns * refclk_mhz;
59 	ret_val /= 1000;
60 
61 	if (ret_val > clamp_value)
62 		ret_val = clamp_value;
63 
64 	return ret_val;
65 }
66 
dcn21_dchvm_init(struct hubbub * hubbub)67 void dcn21_dchvm_init(struct hubbub *hubbub)
68 {
69 	struct dcn20_hubbub *hubbub1 = TO_DCN20_HUBBUB(hubbub);
70 	uint32_t riommu_active;
71 	int i;
72 
73 	//Init DCHVM block
74 	REG_UPDATE(DCHVM_CTRL0, HOSTVM_INIT_REQ, 1);
75 
76 	//Poll until RIOMMU_ACTIVE = 1
77 	for (i = 0; i < 100; i++) {
78 		REG_GET(DCHVM_RIOMMU_STAT0, RIOMMU_ACTIVE, &riommu_active);
79 
80 		if (riommu_active)
81 			break;
82 		else
83 			udelay(5);
84 	}
85 
86 	if (riommu_active) {
87 		//Reflect the power status of DCHUBBUB
88 		REG_UPDATE(DCHVM_RIOMMU_CTRL0, HOSTVM_POWERSTATUS, 1);
89 
90 		//Start rIOMMU prefetching
91 		REG_UPDATE(DCHVM_RIOMMU_CTRL0, HOSTVM_PREFETCH_REQ, 1);
92 
93 		// Enable dynamic clock gating
94 		REG_UPDATE_4(DCHVM_CLK_CTRL,
95 						HVM_DISPCLK_R_GATE_DIS, 0,
96 						HVM_DISPCLK_G_GATE_DIS, 0,
97 						HVM_DCFCLK_R_GATE_DIS, 0,
98 						HVM_DCFCLK_G_GATE_DIS, 0);
99 
100 		//Poll until HOSTVM_PREFETCH_DONE = 1
101 		REG_WAIT(DCHVM_RIOMMU_STAT0, HOSTVM_PREFETCH_DONE, 1, 5, 100);
102 
103 		hubbub->riommu_active = true;
104 	}
105 }
106 
hubbub21_init_dchub(struct hubbub * hubbub,struct dcn_hubbub_phys_addr_config * pa_config)107 int hubbub21_init_dchub(struct hubbub *hubbub,
108 		struct dcn_hubbub_phys_addr_config *pa_config)
109 {
110 	struct dcn20_hubbub *hubbub1 = TO_DCN20_HUBBUB(hubbub);
111 	struct dcn_vmid_page_table_config phys_config;
112 
113 	REG_SET(DCN_VM_FB_LOCATION_BASE, 0,
114 			FB_BASE, pa_config->system_aperture.fb_base >> 24);
115 	REG_SET(DCN_VM_FB_LOCATION_TOP, 0,
116 			FB_TOP, pa_config->system_aperture.fb_top >> 24);
117 	REG_SET(DCN_VM_FB_OFFSET, 0,
118 			FB_OFFSET, pa_config->system_aperture.fb_offset >> 24);
119 	REG_SET(DCN_VM_AGP_BOT, 0,
120 			AGP_BOT, pa_config->system_aperture.agp_bot >> 24);
121 	REG_SET(DCN_VM_AGP_TOP, 0,
122 			AGP_TOP, pa_config->system_aperture.agp_top >> 24);
123 	REG_SET(DCN_VM_AGP_BASE, 0,
124 			AGP_BASE, pa_config->system_aperture.agp_base >> 24);
125 
126 	if (pa_config->gart_config.page_table_start_addr != pa_config->gart_config.page_table_end_addr) {
127 		phys_config.page_table_start_addr = pa_config->gart_config.page_table_start_addr >> 12;
128 		phys_config.page_table_end_addr = pa_config->gart_config.page_table_end_addr >> 12;
129 		phys_config.page_table_base_addr = pa_config->gart_config.page_table_base_addr | 1; //Note: hack
130 		phys_config.depth = 0;
131 		phys_config.block_size = 0;
132 		// Init VMID 0 based on PA config
133 		dcn20_vmid_setup(&hubbub1->vmid[0], &phys_config);
134 	}
135 
136 	dcn21_dchvm_init(hubbub);
137 
138 	return hubbub1->num_vmid;
139 }
140 
hubbub21_program_urgent_watermarks(struct hubbub * hubbub,struct dcn_watermark_set * watermarks,unsigned int refclk_mhz,bool safe_to_lower)141 bool hubbub21_program_urgent_watermarks(
142 		struct hubbub *hubbub,
143 		struct dcn_watermark_set *watermarks,
144 		unsigned int refclk_mhz,
145 		bool safe_to_lower)
146 {
147 	struct dcn20_hubbub *hubbub1 = TO_DCN20_HUBBUB(hubbub);
148 	uint32_t prog_wm_value;
149 	bool wm_pending = false;
150 
151 	/* Repeat for water mark set A, B, C and D. */
152 	/* clock state A */
153 	if (safe_to_lower || watermarks->a.urgent_ns > hubbub1->watermarks.a.urgent_ns) {
154 		hubbub1->watermarks.a.urgent_ns = watermarks->a.urgent_ns;
155 		prog_wm_value = convert_and_clamp(watermarks->a.urgent_ns,
156 				refclk_mhz, 0x1fffff);
157 		REG_SET_2(DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_A, 0,
158 				DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_A, prog_wm_value,
159 				DCHUBBUB_ARB_VM_ROW_URGENCY_WATERMARK_A, prog_wm_value);
160 
161 		DC_LOG_BANDWIDTH_CALCS("URGENCY_WATERMARK_A calculated =%d\n"
162 			"HW register value = 0x%x\n",
163 			watermarks->a.urgent_ns, prog_wm_value);
164 	} else if (watermarks->a.urgent_ns < hubbub1->watermarks.a.urgent_ns)
165 		wm_pending = true;
166 
167 	/* determine the transfer time for a quantity of data for a particular requestor.*/
168 	if (safe_to_lower || watermarks->a.frac_urg_bw_flip
169 			> hubbub1->watermarks.a.frac_urg_bw_flip) {
170 		hubbub1->watermarks.a.frac_urg_bw_flip = watermarks->a.frac_urg_bw_flip;
171 
172 		REG_SET(DCHUBBUB_ARB_FRAC_URG_BW_FLIP_A, 0,
173 				DCHUBBUB_ARB_FRAC_URG_BW_FLIP_A, watermarks->a.frac_urg_bw_flip);
174 	} else if (watermarks->a.frac_urg_bw_flip
175 			< hubbub1->watermarks.a.frac_urg_bw_flip)
176 		wm_pending = true;
177 
178 	if (safe_to_lower || watermarks->a.frac_urg_bw_nom
179 			> hubbub1->watermarks.a.frac_urg_bw_nom) {
180 		hubbub1->watermarks.a.frac_urg_bw_nom = watermarks->a.frac_urg_bw_nom;
181 
182 		REG_SET(DCHUBBUB_ARB_FRAC_URG_BW_NOM_A, 0,
183 				DCHUBBUB_ARB_FRAC_URG_BW_NOM_A, watermarks->a.frac_urg_bw_nom);
184 	} else if (watermarks->a.frac_urg_bw_nom
185 			< hubbub1->watermarks.a.frac_urg_bw_nom)
186 		wm_pending = true;
187 
188 	if (safe_to_lower || watermarks->a.urgent_latency_ns > hubbub1->watermarks.a.urgent_latency_ns) {
189 		hubbub1->watermarks.a.urgent_latency_ns = watermarks->a.urgent_latency_ns;
190 		prog_wm_value = convert_and_clamp(watermarks->a.urgent_latency_ns,
191 				refclk_mhz, 0x1fffff);
192 		REG_SET(DCHUBBUB_ARB_REFCYC_PER_TRIP_TO_MEMORY_A, 0,
193 				DCHUBBUB_ARB_REFCYC_PER_TRIP_TO_MEMORY_A, prog_wm_value);
194 	} else if (watermarks->a.urgent_latency_ns < hubbub1->watermarks.a.urgent_latency_ns)
195 		wm_pending = true;
196 
197 	/* clock state B */
198 	if (safe_to_lower || watermarks->b.urgent_ns > hubbub1->watermarks.b.urgent_ns) {
199 		hubbub1->watermarks.b.urgent_ns = watermarks->b.urgent_ns;
200 		prog_wm_value = convert_and_clamp(watermarks->b.urgent_ns,
201 				refclk_mhz, 0x1fffff);
202 		REG_SET_2(DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_B, 0,
203 				DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_B, prog_wm_value,
204 				DCHUBBUB_ARB_VM_ROW_URGENCY_WATERMARK_B, prog_wm_value);
205 
206 		DC_LOG_BANDWIDTH_CALCS("URGENCY_WATERMARK_B calculated =%d\n"
207 			"HW register value = 0x%x\n",
208 			watermarks->b.urgent_ns, prog_wm_value);
209 	} else if (watermarks->b.urgent_ns < hubbub1->watermarks.b.urgent_ns)
210 		wm_pending = true;
211 
212 	/* determine the transfer time for a quantity of data for a particular requestor.*/
213 	if (safe_to_lower || watermarks->a.frac_urg_bw_flip
214 			> hubbub1->watermarks.a.frac_urg_bw_flip) {
215 		hubbub1->watermarks.a.frac_urg_bw_flip = watermarks->a.frac_urg_bw_flip;
216 
217 		REG_SET(DCHUBBUB_ARB_FRAC_URG_BW_FLIP_B, 0,
218 				DCHUBBUB_ARB_FRAC_URG_BW_FLIP_B, watermarks->a.frac_urg_bw_flip);
219 	} else if (watermarks->a.frac_urg_bw_flip
220 			< hubbub1->watermarks.a.frac_urg_bw_flip)
221 		wm_pending = true;
222 
223 	if (safe_to_lower || watermarks->a.frac_urg_bw_nom
224 			> hubbub1->watermarks.a.frac_urg_bw_nom) {
225 		hubbub1->watermarks.a.frac_urg_bw_nom = watermarks->a.frac_urg_bw_nom;
226 
227 		REG_SET(DCHUBBUB_ARB_FRAC_URG_BW_NOM_B, 0,
228 				DCHUBBUB_ARB_FRAC_URG_BW_NOM_B, watermarks->a.frac_urg_bw_nom);
229 	} else if (watermarks->a.frac_urg_bw_nom
230 			< hubbub1->watermarks.a.frac_urg_bw_nom)
231 		wm_pending = true;
232 
233 	if (safe_to_lower || watermarks->b.urgent_latency_ns > hubbub1->watermarks.b.urgent_latency_ns) {
234 		hubbub1->watermarks.b.urgent_latency_ns = watermarks->b.urgent_latency_ns;
235 		prog_wm_value = convert_and_clamp(watermarks->b.urgent_latency_ns,
236 				refclk_mhz, 0x1fffff);
237 		REG_SET(DCHUBBUB_ARB_REFCYC_PER_TRIP_TO_MEMORY_B, 0,
238 				DCHUBBUB_ARB_REFCYC_PER_TRIP_TO_MEMORY_B, prog_wm_value);
239 	} else if (watermarks->b.urgent_latency_ns < hubbub1->watermarks.b.urgent_latency_ns)
240 		wm_pending = true;
241 
242 	/* clock state C */
243 	if (safe_to_lower || watermarks->c.urgent_ns > hubbub1->watermarks.c.urgent_ns) {
244 		hubbub1->watermarks.c.urgent_ns = watermarks->c.urgent_ns;
245 		prog_wm_value = convert_and_clamp(watermarks->c.urgent_ns,
246 				refclk_mhz, 0x1fffff);
247 		REG_SET_2(DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_C, 0,
248 				DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_C, prog_wm_value,
249 				DCHUBBUB_ARB_VM_ROW_URGENCY_WATERMARK_C, prog_wm_value);
250 
251 		DC_LOG_BANDWIDTH_CALCS("URGENCY_WATERMARK_C calculated =%d\n"
252 			"HW register value = 0x%x\n",
253 			watermarks->c.urgent_ns, prog_wm_value);
254 	} else if (watermarks->c.urgent_ns < hubbub1->watermarks.c.urgent_ns)
255 		wm_pending = true;
256 
257 	/* determine the transfer time for a quantity of data for a particular requestor.*/
258 	if (safe_to_lower || watermarks->a.frac_urg_bw_flip
259 			> hubbub1->watermarks.a.frac_urg_bw_flip) {
260 		hubbub1->watermarks.a.frac_urg_bw_flip = watermarks->a.frac_urg_bw_flip;
261 
262 		REG_SET(DCHUBBUB_ARB_FRAC_URG_BW_FLIP_C, 0,
263 				DCHUBBUB_ARB_FRAC_URG_BW_FLIP_C, watermarks->a.frac_urg_bw_flip);
264 	} else if (watermarks->a.frac_urg_bw_flip
265 			< hubbub1->watermarks.a.frac_urg_bw_flip)
266 		wm_pending = true;
267 
268 	if (safe_to_lower || watermarks->a.frac_urg_bw_nom
269 			> hubbub1->watermarks.a.frac_urg_bw_nom) {
270 		hubbub1->watermarks.a.frac_urg_bw_nom = watermarks->a.frac_urg_bw_nom;
271 
272 		REG_SET(DCHUBBUB_ARB_FRAC_URG_BW_NOM_C, 0,
273 				DCHUBBUB_ARB_FRAC_URG_BW_NOM_C, watermarks->a.frac_urg_bw_nom);
274 	} else if (watermarks->a.frac_urg_bw_nom
275 			< hubbub1->watermarks.a.frac_urg_bw_nom)
276 		wm_pending = true;
277 
278 	if (safe_to_lower || watermarks->c.urgent_latency_ns > hubbub1->watermarks.c.urgent_latency_ns) {
279 		hubbub1->watermarks.c.urgent_latency_ns = watermarks->c.urgent_latency_ns;
280 		prog_wm_value = convert_and_clamp(watermarks->c.urgent_latency_ns,
281 				refclk_mhz, 0x1fffff);
282 		REG_SET(DCHUBBUB_ARB_REFCYC_PER_TRIP_TO_MEMORY_C, 0,
283 				DCHUBBUB_ARB_REFCYC_PER_TRIP_TO_MEMORY_C, prog_wm_value);
284 	} else if (watermarks->c.urgent_latency_ns < hubbub1->watermarks.c.urgent_latency_ns)
285 		wm_pending = true;
286 
287 	/* clock state D */
288 	if (safe_to_lower || watermarks->d.urgent_ns > hubbub1->watermarks.d.urgent_ns) {
289 		hubbub1->watermarks.d.urgent_ns = watermarks->d.urgent_ns;
290 		prog_wm_value = convert_and_clamp(watermarks->d.urgent_ns,
291 				refclk_mhz, 0x1fffff);
292 		REG_SET_2(DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_D, 0,
293 				DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_D, prog_wm_value,
294 				DCHUBBUB_ARB_VM_ROW_URGENCY_WATERMARK_D, prog_wm_value);
295 
296 		DC_LOG_BANDWIDTH_CALCS("URGENCY_WATERMARK_D calculated =%d\n"
297 			"HW register value = 0x%x\n",
298 			watermarks->d.urgent_ns, prog_wm_value);
299 	} else if (watermarks->d.urgent_ns < hubbub1->watermarks.d.urgent_ns)
300 		wm_pending = true;
301 
302 	/* determine the transfer time for a quantity of data for a particular requestor.*/
303 	if (safe_to_lower || watermarks->a.frac_urg_bw_flip
304 			> hubbub1->watermarks.a.frac_urg_bw_flip) {
305 		hubbub1->watermarks.a.frac_urg_bw_flip = watermarks->a.frac_urg_bw_flip;
306 
307 		REG_SET(DCHUBBUB_ARB_FRAC_URG_BW_FLIP_D, 0,
308 				DCHUBBUB_ARB_FRAC_URG_BW_FLIP_D, watermarks->a.frac_urg_bw_flip);
309 	} else if (watermarks->a.frac_urg_bw_flip
310 			< hubbub1->watermarks.a.frac_urg_bw_flip)
311 		wm_pending = true;
312 
313 	if (safe_to_lower || watermarks->a.frac_urg_bw_nom
314 			> hubbub1->watermarks.a.frac_urg_bw_nom) {
315 		hubbub1->watermarks.a.frac_urg_bw_nom = watermarks->a.frac_urg_bw_nom;
316 
317 		REG_SET(DCHUBBUB_ARB_FRAC_URG_BW_NOM_D, 0,
318 				DCHUBBUB_ARB_FRAC_URG_BW_NOM_D, watermarks->a.frac_urg_bw_nom);
319 	} else if (watermarks->a.frac_urg_bw_nom
320 			< hubbub1->watermarks.a.frac_urg_bw_nom)
321 		wm_pending = true;
322 
323 	if (safe_to_lower || watermarks->d.urgent_latency_ns > hubbub1->watermarks.d.urgent_latency_ns) {
324 		hubbub1->watermarks.d.urgent_latency_ns = watermarks->d.urgent_latency_ns;
325 		prog_wm_value = convert_and_clamp(watermarks->d.urgent_latency_ns,
326 				refclk_mhz, 0x1fffff);
327 		REG_SET(DCHUBBUB_ARB_REFCYC_PER_TRIP_TO_MEMORY_D, 0,
328 				DCHUBBUB_ARB_REFCYC_PER_TRIP_TO_MEMORY_D, prog_wm_value);
329 	} else if (watermarks->d.urgent_latency_ns < hubbub1->watermarks.d.urgent_latency_ns)
330 		wm_pending = true;
331 
332 	return wm_pending;
333 }
334 
hubbub21_program_stutter_watermarks(struct hubbub * hubbub,struct dcn_watermark_set * watermarks,unsigned int refclk_mhz,bool safe_to_lower)335 bool hubbub21_program_stutter_watermarks(
336 		struct hubbub *hubbub,
337 		struct dcn_watermark_set *watermarks,
338 		unsigned int refclk_mhz,
339 		bool safe_to_lower)
340 {
341 	struct dcn20_hubbub *hubbub1 = TO_DCN20_HUBBUB(hubbub);
342 	uint32_t prog_wm_value;
343 	bool wm_pending = false;
344 
345 	/* clock state A */
346 	if (safe_to_lower || watermarks->a.cstate_pstate.cstate_enter_plus_exit_ns
347 			> hubbub1->watermarks.a.cstate_pstate.cstate_enter_plus_exit_ns) {
348 		hubbub1->watermarks.a.cstate_pstate.cstate_enter_plus_exit_ns =
349 				watermarks->a.cstate_pstate.cstate_enter_plus_exit_ns;
350 		prog_wm_value = convert_and_clamp(
351 				watermarks->a.cstate_pstate.cstate_enter_plus_exit_ns,
352 				refclk_mhz, 0x1fffff);
353 		REG_SET_2(DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_A, 0,
354 				DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_A, prog_wm_value,
355 				DCHUBBUB_ARB_VM_ROW_ALLOW_SR_ENTER_WATERMARK_A, prog_wm_value);
356 		DC_LOG_BANDWIDTH_CALCS("SR_ENTER_EXIT_WATERMARK_A calculated =%d\n"
357 			"HW register value = 0x%x\n",
358 			watermarks->a.cstate_pstate.cstate_enter_plus_exit_ns, prog_wm_value);
359 	} else if (watermarks->a.cstate_pstate.cstate_enter_plus_exit_ns
360 			< hubbub1->watermarks.a.cstate_pstate.cstate_enter_plus_exit_ns)
361 		wm_pending = true;
362 
363 	if (safe_to_lower || watermarks->a.cstate_pstate.cstate_exit_ns
364 			> hubbub1->watermarks.a.cstate_pstate.cstate_exit_ns) {
365 		hubbub1->watermarks.a.cstate_pstate.cstate_exit_ns =
366 				watermarks->a.cstate_pstate.cstate_exit_ns;
367 		prog_wm_value = convert_and_clamp(
368 				watermarks->a.cstate_pstate.cstate_exit_ns,
369 				refclk_mhz, 0x1fffff);
370 		REG_SET_2(DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_A, 0,
371 				DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_A, prog_wm_value,
372 				DCHUBBUB_ARB_VM_ROW_ALLOW_SR_EXIT_WATERMARK_A, prog_wm_value);
373 		DC_LOG_BANDWIDTH_CALCS("SR_EXIT_WATERMARK_A calculated =%d\n"
374 			"HW register value = 0x%x\n",
375 			watermarks->a.cstate_pstate.cstate_exit_ns, prog_wm_value);
376 	} else if (watermarks->a.cstate_pstate.cstate_exit_ns
377 			< hubbub1->watermarks.a.cstate_pstate.cstate_exit_ns)
378 		wm_pending = true;
379 
380 	/* clock state B */
381 	if (safe_to_lower || watermarks->b.cstate_pstate.cstate_enter_plus_exit_ns
382 			> hubbub1->watermarks.b.cstate_pstate.cstate_enter_plus_exit_ns) {
383 		hubbub1->watermarks.b.cstate_pstate.cstate_enter_plus_exit_ns =
384 				watermarks->b.cstate_pstate.cstate_enter_plus_exit_ns;
385 		prog_wm_value = convert_and_clamp(
386 				watermarks->b.cstate_pstate.cstate_enter_plus_exit_ns,
387 				refclk_mhz, 0x1fffff);
388 		REG_SET_2(DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_B, 0,
389 				DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_B, prog_wm_value,
390 				DCHUBBUB_ARB_VM_ROW_ALLOW_SR_ENTER_WATERMARK_B, prog_wm_value);
391 		DC_LOG_BANDWIDTH_CALCS("SR_ENTER_EXIT_WATERMARK_B calculated =%d\n"
392 			"HW register value = 0x%x\n",
393 			watermarks->b.cstate_pstate.cstate_enter_plus_exit_ns, prog_wm_value);
394 	} else if (watermarks->b.cstate_pstate.cstate_enter_plus_exit_ns
395 			< hubbub1->watermarks.b.cstate_pstate.cstate_enter_plus_exit_ns)
396 		wm_pending = true;
397 
398 	if (safe_to_lower || watermarks->b.cstate_pstate.cstate_exit_ns
399 			> hubbub1->watermarks.b.cstate_pstate.cstate_exit_ns) {
400 		hubbub1->watermarks.b.cstate_pstate.cstate_exit_ns =
401 				watermarks->b.cstate_pstate.cstate_exit_ns;
402 		prog_wm_value = convert_and_clamp(
403 				watermarks->b.cstate_pstate.cstate_exit_ns,
404 				refclk_mhz, 0x1fffff);
405 		REG_SET_2(DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_B, 0,
406 				DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_B, prog_wm_value,
407 				DCHUBBUB_ARB_VM_ROW_ALLOW_SR_EXIT_WATERMARK_A, prog_wm_value);
408 		DC_LOG_BANDWIDTH_CALCS("SR_EXIT_WATERMARK_B calculated =%d\n"
409 			"HW register value = 0x%x\n",
410 			watermarks->b.cstate_pstate.cstate_exit_ns, prog_wm_value);
411 	} else if (watermarks->b.cstate_pstate.cstate_exit_ns
412 			< hubbub1->watermarks.b.cstate_pstate.cstate_exit_ns)
413 		wm_pending = true;
414 
415 	/* clock state C */
416 	if (safe_to_lower || watermarks->c.cstate_pstate.cstate_enter_plus_exit_ns
417 			> hubbub1->watermarks.c.cstate_pstate.cstate_enter_plus_exit_ns) {
418 		hubbub1->watermarks.c.cstate_pstate.cstate_enter_plus_exit_ns =
419 				watermarks->c.cstate_pstate.cstate_enter_plus_exit_ns;
420 		prog_wm_value = convert_and_clamp(
421 				watermarks->c.cstate_pstate.cstate_enter_plus_exit_ns,
422 				refclk_mhz, 0x1fffff);
423 		REG_SET_2(DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_C, 0,
424 				DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_C, prog_wm_value,
425 				DCHUBBUB_ARB_VM_ROW_ALLOW_SR_ENTER_WATERMARK_C, prog_wm_value);
426 		DC_LOG_BANDWIDTH_CALCS("SR_ENTER_EXIT_WATERMARK_C calculated =%d\n"
427 			"HW register value = 0x%x\n",
428 			watermarks->c.cstate_pstate.cstate_enter_plus_exit_ns, prog_wm_value);
429 	} else if (watermarks->c.cstate_pstate.cstate_enter_plus_exit_ns
430 			< hubbub1->watermarks.c.cstate_pstate.cstate_enter_plus_exit_ns)
431 		wm_pending = true;
432 
433 	if (safe_to_lower || watermarks->c.cstate_pstate.cstate_exit_ns
434 			> hubbub1->watermarks.c.cstate_pstate.cstate_exit_ns) {
435 		hubbub1->watermarks.c.cstate_pstate.cstate_exit_ns =
436 				watermarks->c.cstate_pstate.cstate_exit_ns;
437 		prog_wm_value = convert_and_clamp(
438 				watermarks->c.cstate_pstate.cstate_exit_ns,
439 				refclk_mhz, 0x1fffff);
440 		REG_SET_2(DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_C, 0,
441 				DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_C, prog_wm_value,
442 				DCHUBBUB_ARB_VM_ROW_ALLOW_SR_EXIT_WATERMARK_A, prog_wm_value);
443 		DC_LOG_BANDWIDTH_CALCS("SR_EXIT_WATERMARK_C calculated =%d\n"
444 			"HW register value = 0x%x\n",
445 			watermarks->c.cstate_pstate.cstate_exit_ns, prog_wm_value);
446 	} else if (watermarks->c.cstate_pstate.cstate_exit_ns
447 			< hubbub1->watermarks.c.cstate_pstate.cstate_exit_ns)
448 		wm_pending = true;
449 
450 	/* clock state D */
451 	if (safe_to_lower || watermarks->d.cstate_pstate.cstate_enter_plus_exit_ns
452 			> hubbub1->watermarks.d.cstate_pstate.cstate_enter_plus_exit_ns) {
453 		hubbub1->watermarks.d.cstate_pstate.cstate_enter_plus_exit_ns =
454 				watermarks->d.cstate_pstate.cstate_enter_plus_exit_ns;
455 		prog_wm_value = convert_and_clamp(
456 				watermarks->d.cstate_pstate.cstate_enter_plus_exit_ns,
457 				refclk_mhz, 0x1fffff);
458 		REG_SET_2(DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_D, 0,
459 				DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_D, prog_wm_value,
460 				DCHUBBUB_ARB_VM_ROW_ALLOW_SR_ENTER_WATERMARK_D, prog_wm_value);
461 		DC_LOG_BANDWIDTH_CALCS("SR_ENTER_EXIT_WATERMARK_D calculated =%d\n"
462 			"HW register value = 0x%x\n",
463 			watermarks->d.cstate_pstate.cstate_enter_plus_exit_ns, prog_wm_value);
464 	} else if (watermarks->d.cstate_pstate.cstate_enter_plus_exit_ns
465 			< hubbub1->watermarks.d.cstate_pstate.cstate_enter_plus_exit_ns)
466 		wm_pending = true;
467 
468 	if (safe_to_lower || watermarks->d.cstate_pstate.cstate_exit_ns
469 			> hubbub1->watermarks.d.cstate_pstate.cstate_exit_ns) {
470 		hubbub1->watermarks.d.cstate_pstate.cstate_exit_ns =
471 				watermarks->d.cstate_pstate.cstate_exit_ns;
472 		prog_wm_value = convert_and_clamp(
473 				watermarks->d.cstate_pstate.cstate_exit_ns,
474 				refclk_mhz, 0x1fffff);
475 		REG_SET_2(DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_D, 0,
476 				DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_D, prog_wm_value,
477 				DCHUBBUB_ARB_VM_ROW_ALLOW_SR_EXIT_WATERMARK_A, prog_wm_value);
478 		DC_LOG_BANDWIDTH_CALCS("SR_EXIT_WATERMARK_D calculated =%d\n"
479 			"HW register value = 0x%x\n",
480 			watermarks->d.cstate_pstate.cstate_exit_ns, prog_wm_value);
481 	} else if (watermarks->d.cstate_pstate.cstate_exit_ns
482 			< hubbub1->watermarks.d.cstate_pstate.cstate_exit_ns)
483 		wm_pending = true;
484 
485 	return wm_pending;
486 }
487 
hubbub21_program_pstate_watermarks(struct hubbub * hubbub,struct dcn_watermark_set * watermarks,unsigned int refclk_mhz,bool safe_to_lower)488 bool hubbub21_program_pstate_watermarks(
489 		struct hubbub *hubbub,
490 		struct dcn_watermark_set *watermarks,
491 		unsigned int refclk_mhz,
492 		bool safe_to_lower)
493 {
494 	struct dcn20_hubbub *hubbub1 = TO_DCN20_HUBBUB(hubbub);
495 	uint32_t prog_wm_value;
496 
497 	bool wm_pending = false;
498 
499 	/* clock state A */
500 	if (safe_to_lower || watermarks->a.cstate_pstate.pstate_change_ns
501 			> hubbub1->watermarks.a.cstate_pstate.pstate_change_ns) {
502 		hubbub1->watermarks.a.cstate_pstate.pstate_change_ns =
503 				watermarks->a.cstate_pstate.pstate_change_ns;
504 		prog_wm_value = convert_and_clamp(
505 				watermarks->a.cstate_pstate.pstate_change_ns,
506 				refclk_mhz, 0x1fffff);
507 		REG_SET_2(DCHUBBUB_ARB_ALLOW_DRAM_CLK_CHANGE_WATERMARK_A, 0,
508 				DCHUBBUB_ARB_ALLOW_DRAM_CLK_CHANGE_WATERMARK_A, prog_wm_value,
509 				DCHUBBUB_ARB_VM_ROW_ALLOW_DRAM_CLK_CHANGE_WATERMARK_A, prog_wm_value);
510 		DC_LOG_BANDWIDTH_CALCS("DRAM_CLK_CHANGE_WATERMARK_A calculated =%d\n"
511 			"HW register value = 0x%x\n\n",
512 			watermarks->a.cstate_pstate.pstate_change_ns, prog_wm_value);
513 	} else if (watermarks->a.cstate_pstate.pstate_change_ns
514 			< hubbub1->watermarks.a.cstate_pstate.pstate_change_ns)
515 		wm_pending = true;
516 
517 	/* clock state B */
518 	if (safe_to_lower || watermarks->b.cstate_pstate.pstate_change_ns
519 			> hubbub1->watermarks.b.cstate_pstate.pstate_change_ns) {
520 		hubbub1->watermarks.b.cstate_pstate.pstate_change_ns =
521 				watermarks->b.cstate_pstate.pstate_change_ns;
522 		prog_wm_value = convert_and_clamp(
523 				watermarks->b.cstate_pstate.pstate_change_ns,
524 				refclk_mhz, 0x1fffff);
525 		REG_SET_2(DCHUBBUB_ARB_ALLOW_DRAM_CLK_CHANGE_WATERMARK_B, 0,
526 				DCHUBBUB_ARB_ALLOW_DRAM_CLK_CHANGE_WATERMARK_B, prog_wm_value,
527 				DCHUBBUB_ARB_VM_ROW_ALLOW_DRAM_CLK_CHANGE_WATERMARK_B, prog_wm_value);
528 		DC_LOG_BANDWIDTH_CALCS("DRAM_CLK_CHANGE_WATERMARK_B calculated =%d\n"
529 			"HW register value = 0x%x\n\n",
530 			watermarks->b.cstate_pstate.pstate_change_ns, prog_wm_value);
531 	} else if (watermarks->b.cstate_pstate.pstate_change_ns
532 			< hubbub1->watermarks.b.cstate_pstate.pstate_change_ns)
533 		wm_pending = false;
534 
535 	/* clock state C */
536 	if (safe_to_lower || watermarks->c.cstate_pstate.pstate_change_ns
537 			> hubbub1->watermarks.c.cstate_pstate.pstate_change_ns) {
538 		hubbub1->watermarks.c.cstate_pstate.pstate_change_ns =
539 				watermarks->c.cstate_pstate.pstate_change_ns;
540 		prog_wm_value = convert_and_clamp(
541 				watermarks->c.cstate_pstate.pstate_change_ns,
542 				refclk_mhz, 0x1fffff);
543 		REG_SET_2(DCHUBBUB_ARB_ALLOW_DRAM_CLK_CHANGE_WATERMARK_C, 0,
544 				DCHUBBUB_ARB_ALLOW_DRAM_CLK_CHANGE_WATERMARK_C, prog_wm_value,
545 				DCHUBBUB_ARB_VM_ROW_ALLOW_DRAM_CLK_CHANGE_WATERMARK_C, prog_wm_value);
546 		DC_LOG_BANDWIDTH_CALCS("DRAM_CLK_CHANGE_WATERMARK_C calculated =%d\n"
547 			"HW register value = 0x%x\n\n",
548 			watermarks->c.cstate_pstate.pstate_change_ns, prog_wm_value);
549 	} else if (watermarks->c.cstate_pstate.pstate_change_ns
550 			< hubbub1->watermarks.c.cstate_pstate.pstate_change_ns)
551 		wm_pending = true;
552 
553 	/* clock state D */
554 	if (safe_to_lower || watermarks->d.cstate_pstate.pstate_change_ns
555 			> hubbub1->watermarks.d.cstate_pstate.pstate_change_ns) {
556 		hubbub1->watermarks.d.cstate_pstate.pstate_change_ns =
557 				watermarks->d.cstate_pstate.pstate_change_ns;
558 		prog_wm_value = convert_and_clamp(
559 				watermarks->d.cstate_pstate.pstate_change_ns,
560 				refclk_mhz, 0x1fffff);
561 		REG_SET_2(DCHUBBUB_ARB_ALLOW_DRAM_CLK_CHANGE_WATERMARK_D, 0,
562 				DCHUBBUB_ARB_ALLOW_DRAM_CLK_CHANGE_WATERMARK_D, prog_wm_value,
563 				DCHUBBUB_ARB_VM_ROW_ALLOW_DRAM_CLK_CHANGE_WATERMARK_D, prog_wm_value);
564 		DC_LOG_BANDWIDTH_CALCS("DRAM_CLK_CHANGE_WATERMARK_D calculated =%d\n"
565 			"HW register value = 0x%x\n\n",
566 			watermarks->d.cstate_pstate.pstate_change_ns, prog_wm_value);
567 	} else if (watermarks->d.cstate_pstate.pstate_change_ns
568 			< hubbub1->watermarks.d.cstate_pstate.pstate_change_ns)
569 		wm_pending = true;
570 
571 	return wm_pending;
572 }
573 
hubbub21_program_watermarks(struct hubbub * hubbub,struct dcn_watermark_set * watermarks,unsigned int refclk_mhz,bool safe_to_lower)574 bool hubbub21_program_watermarks(
575 		struct hubbub *hubbub,
576 		struct dcn_watermark_set *watermarks,
577 		unsigned int refclk_mhz,
578 		bool safe_to_lower)
579 {
580 	struct dcn20_hubbub *hubbub1 = TO_DCN20_HUBBUB(hubbub);
581 	bool wm_pending = false;
582 
583 	if (hubbub21_program_urgent_watermarks(hubbub, watermarks, refclk_mhz, safe_to_lower))
584 		wm_pending = true;
585 
586 	if (hubbub21_program_stutter_watermarks(hubbub, watermarks, refclk_mhz, safe_to_lower))
587 		wm_pending = true;
588 
589 	if (hubbub21_program_pstate_watermarks(hubbub, watermarks, refclk_mhz, safe_to_lower))
590 		wm_pending = true;
591 
592 	/*
593 	 * The DCHub arbiter has a mechanism to dynamically rate limit the DCHub request stream to the fabric.
594 	 * If the memory controller is fully utilized and the DCHub requestors are
595 	 * well ahead of their amortized schedule, then it is safe to prevent the next winner
596 	 * from being committed and sent to the fabric.
597 	 * The utilization of the memory controller is approximated by ensuring that
598 	 * the number of outstanding requests is greater than a threshold specified
599 	 * by the ARB_MIN_REQ_OUTSTANDING. To determine that the DCHub requestors are well ahead of the amortized schedule,
600 	 * the slack of the next winner is compared with the ARB_SAT_LEVEL in DLG RefClk cycles.
601 	 *
602 	 * TODO: Revisit request limit after figure out right number. request limit for Renoir isn't decided yet, set maximum value (0x1FF)
603 	 * to turn off it for now.
604 	 */
605 	REG_SET(DCHUBBUB_ARB_SAT_LEVEL, 0,
606 			DCHUBBUB_ARB_SAT_LEVEL, 60 * refclk_mhz);
607 	REG_UPDATE_2(DCHUBBUB_ARB_DF_REQ_OUTSTAND,
608 			DCHUBBUB_ARB_MIN_REQ_OUTSTAND, 0x1FF,
609 			DCHUBBUB_ARB_MIN_REQ_OUTSTAND_COMMIT_THRESHOLD, 0xA);
610 	REG_UPDATE(DCHUBBUB_ARB_HOSTVM_CNTL,
611 			DCHUBBUB_ARB_MAX_QOS_COMMIT_THRESHOLD, 0xF);
612 
613 	hubbub1_allow_self_refresh_control(hubbub, !hubbub->ctx->dc->debug.disable_stutter);
614 
615 	return wm_pending;
616 }
617 
hubbub21_wm_read_state(struct hubbub * hubbub,struct dcn_hubbub_wm * wm)618 void hubbub21_wm_read_state(struct hubbub *hubbub,
619 		struct dcn_hubbub_wm *wm)
620 {
621 	struct dcn20_hubbub *hubbub1 = TO_DCN20_HUBBUB(hubbub);
622 	struct dcn_hubbub_wm_set *s;
623 
624 	memset(wm, 0, sizeof(struct dcn_hubbub_wm));
625 
626 	s = &wm->sets[0];
627 	s->wm_set = 0;
628 	REG_GET(DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_A,
629 			DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_A, &s->data_urgent);
630 
631 	REG_GET(DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_A,
632 			DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_A, &s->sr_enter);
633 
634 	REG_GET(DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_A,
635 			DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_A, &s->sr_exit);
636 
637 	REG_GET(DCHUBBUB_ARB_ALLOW_DRAM_CLK_CHANGE_WATERMARK_A,
638 			 DCHUBBUB_ARB_ALLOW_DRAM_CLK_CHANGE_WATERMARK_A, &s->dram_clk_chanage);
639 
640 	s = &wm->sets[1];
641 	s->wm_set = 1;
642 	REG_GET(DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_B,
643 			DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_B, &s->data_urgent);
644 
645 	REG_GET(DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_B,
646 			DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_B, &s->sr_enter);
647 
648 	REG_GET(DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_B,
649 			DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_B, &s->sr_exit);
650 
651 	REG_GET(DCHUBBUB_ARB_ALLOW_DRAM_CLK_CHANGE_WATERMARK_B,
652 			DCHUBBUB_ARB_ALLOW_DRAM_CLK_CHANGE_WATERMARK_B, &s->dram_clk_chanage);
653 
654 	s = &wm->sets[2];
655 	s->wm_set = 2;
656 	REG_GET(DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_C,
657 			DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_C, &s->data_urgent);
658 
659 	REG_GET(DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_C,
660 			DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_C, &s->sr_enter);
661 
662 	REG_GET(DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_C,
663 			DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_C, &s->sr_exit);
664 
665 	REG_GET(DCHUBBUB_ARB_ALLOW_DRAM_CLK_CHANGE_WATERMARK_C,
666 			DCHUBBUB_ARB_ALLOW_DRAM_CLK_CHANGE_WATERMARK_C, &s->dram_clk_chanage);
667 
668 	s = &wm->sets[3];
669 	s->wm_set = 3;
670 	REG_GET(DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_D,
671 			DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_D, &s->data_urgent);
672 
673 	REG_GET(DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_D,
674 			DCHUBBUB_ARB_ALLOW_SR_ENTER_WATERMARK_D, &s->sr_enter);
675 
676 	REG_GET(DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_D,
677 			DCHUBBUB_ARB_ALLOW_SR_EXIT_WATERMARK_D, &s->sr_exit);
678 
679 	REG_GET(DCHUBBUB_ARB_ALLOW_DRAM_CLK_CHANGE_WATERMARK_D,
680 			DCHUBBUB_ARB_ALLOW_DRAM_CLK_CHANGE_WATERMARK_D, &s->dram_clk_chanage);
681 }
682 
hubbub21_apply_DEDCN21_147_wa(struct hubbub * hubbub)683 static void hubbub21_apply_DEDCN21_147_wa(struct hubbub *hubbub)
684 {
685 	struct dcn20_hubbub *hubbub1 = TO_DCN20_HUBBUB(hubbub);
686 	uint32_t prog_wm_value;
687 
688 	prog_wm_value = REG_READ(DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_A);
689 	REG_WRITE(DCHUBBUB_ARB_DATA_URGENCY_WATERMARK_A, prog_wm_value);
690 }
691 
692 static const struct hubbub_funcs hubbub21_funcs = {
693 	.update_dchub = hubbub2_update_dchub,
694 	.init_dchub_sys_ctx = hubbub21_init_dchub,
695 	.init_vm_ctx = hubbub2_init_vm_ctx,
696 	.dcc_support_swizzle = hubbub2_dcc_support_swizzle,
697 	.dcc_support_pixel_format = hubbub2_dcc_support_pixel_format,
698 	.get_dcc_compression_cap = hubbub2_get_dcc_compression_cap,
699 	.wm_read_state = hubbub21_wm_read_state,
700 	.get_dchub_ref_freq = hubbub2_get_dchub_ref_freq,
701 	.program_watermarks = hubbub21_program_watermarks,
702 	.allow_self_refresh_control = hubbub1_allow_self_refresh_control,
703 	.apply_DEDCN21_147_wa = hubbub21_apply_DEDCN21_147_wa,
704 	.hubbub_read_state = hubbub2_read_state,
705 };
706 
hubbub21_construct(struct dcn20_hubbub * hubbub,struct dc_context * ctx,const struct dcn_hubbub_registers * hubbub_regs,const struct dcn_hubbub_shift * hubbub_shift,const struct dcn_hubbub_mask * hubbub_mask)707 void hubbub21_construct(struct dcn20_hubbub *hubbub,
708 	struct dc_context *ctx,
709 	const struct dcn_hubbub_registers *hubbub_regs,
710 	const struct dcn_hubbub_shift *hubbub_shift,
711 	const struct dcn_hubbub_mask *hubbub_mask)
712 {
713 	hubbub->base.ctx = ctx;
714 
715 	hubbub->base.funcs = &hubbub21_funcs;
716 
717 	hubbub->regs = hubbub_regs;
718 	hubbub->shifts = hubbub_shift;
719 	hubbub->masks = hubbub_mask;
720 
721 	hubbub->debug_test_index_pstate = 0xB;
722 	hubbub->detile_buf_size = 164 * 1024; /* 164KB for DCN2.0 */
723 }
724