From 653e25bb3b8897288bb13bad4ca999fedfb28a3e Mon Sep 17 00:00:00 2001 From: Pepping <pepping> Date: Mon, 26 Jan 2015 14:13:42 +0000 Subject: [PATCH] - Added c_force_late_sync and c_force_early_sync - Fixed wrong indent of ref_data_total --- .../reorder/tb/python/tc_reorder_transpose.py | 41 ++++++++++--------- 1 file changed, 21 insertions(+), 20 deletions(-) diff --git a/libraries/base/reorder/tb/python/tc_reorder_transpose.py b/libraries/base/reorder/tb/python/tc_reorder_transpose.py index d08397e391..d5ebd7e144 100644 --- a/libraries/base/reorder/tb/python/tc_reorder_transpose.py +++ b/libraries/base/reorder/tb/python/tc_reorder_transpose.py @@ -88,8 +88,11 @@ c_db_nof_streams = c_bg_nof_streams c_db_ram_size = c_bg_ram_size #g_rd_chunksize * g_rd_nof_chunks * g_nof_blocks c_frame_size = g_wr_chunksize c_nof_int_streams = 1 -c_ena_pre_transpose = False +c_ena_pre_transpose = True c_gap_size = 0 #g_rd_chunksize +c_force_late_sync = 0 +c_force_early_sync = 0 + tc.append_log(3, '>>>') tc.append_log(1, '>>> Title : Test bench for ddr3_transpose' ) @@ -121,7 +124,7 @@ def gen_bg_hex_files(c_nof_values = 1024, c_nof_streams = 4): stream_im = [] for j in range(c_nof_values): stream_re.append(j) - stream_im.append(i) + stream_im.append(i) data_concat = dsp_test_bg.concatenate_two_lists(stream_re, stream_im, c_in_dat_w) data.append(data_concat) filename = "../../src/hex/tb_bg_dat_" + str(i) + ".hex" @@ -139,11 +142,11 @@ if __name__ == "__main__": for j in range(g_rd_chunksize): ss_list.append(i + j*c_frame_size) - + if c_ena_pre_transpose: + ss.write_selects(ss_list) + # for i in ss_list: -# print i -# ss.write_selects(ss_list) - +# print i ############################################################################### # # Create stimuli for the BG @@ -179,12 +182,17 @@ if __name__ == "__main__": do_until_gt(io.simIO.getSimTime, ms_retry=1000, val=110000, s_timeout=13600) # 110000 # Enable the blockgenerator - bg.write_enable() - do_until_gt(io.simIO.getSimTime, ms_retry=1000, val=180000, s_timeout=13600) # 110000 - bg.write_disable() + bg.write_enable() - do_until_gt(io.simIO.getSimTime, ms_retry=1000, val=182000, s_timeout=13600) # 110000 - bg.write_enable() + if(c_force_late_sync == 1): + do_until_gt(io.simIO.getSimTime, ms_retry=1000, val=180000, s_timeout=13600) # 110000 + bg.write_block_gen_settings(samplesPerPacket=c_frame_size, blocksPerSync=g_nof_blk_per_sync+1, gapSize=c_gap_size, memLowAddr=0, memHighAddr=c_bg_ram_size-1, BSNInit=10) + elif(c_force_early_sync == 1): + do_until_gt(io.simIO.getSimTime, ms_retry=1000, val=180000, s_timeout=13600) # 110000 + bg.write_block_gen_settings(samplesPerPacket=c_frame_size, blocksPerSync=g_nof_blk_per_sync-1, gapSize=c_gap_size, memLowAddr=0, memHighAddr=c_bg_ram_size-1, BSNInit=10) + + #do_until_gt(io.simIO.getSimTime, ms_retry=1000, val=182000, s_timeout=13600) # 110000 + #bg.write_enable() # #do_until_gt(io.simIO.getSimTime, ms_retry=1000, val=120000, s_timeout=13600) # 110000 #bg.write_disable() @@ -197,8 +205,6 @@ if __name__ == "__main__": # ############################################################################### # Subband Select pre-transpose - print "len(ss_list)" - print len(ss_list) if c_ena_pre_transpose: bg_data = ss.subband_select(bg_data, ss_list) @@ -208,11 +214,6 @@ if __name__ == "__main__": bg_data_single_page = bg_data[t*c_pagesize:(t+1)*c_pagesize] # Write to memory mem_page = [0] * c_pagesize - print "len(mem_page)" - print len(mem_page) - print "len(bg_data_single_page)" - print len(bg_data_single_page) - for i in range(g_nof_blocks): for j in range(g_wr_nof_chunks): for k in range(g_wr_chunksize): @@ -231,10 +232,10 @@ if __name__ == "__main__": #ref_data[chunk_cnt*(g_rd_chunksize*g_rd_nof_chunks)+ chunk_offset*g_rd_chunksize + k] = mem_page[chunk_cnt*(g_rd_chunksize*g_rd_nof_chunks+g_gapsize)+ chunk_offset*g_rd_chunksize + k] ref_data[i*(g_rd_chunksize*g_rd_nof_chunks)+j*g_rd_chunksize + k] = mem_page[chunk_cnt*(g_rd_chunksize*g_rd_nof_chunks+g_gapsize)+ chunk_offset*g_rd_chunksize + k] chunk_cnt = chunk_cnt + 1 - ref_data_total.append(ref_data) + ref_data_total.append(ref_data) ref_data_total=flatten(ref_data_total) - + # Split the data again in individual channels ref_data_split = [] ref_data_split.append(ref_data_total) -- GitLab