From 653e25bb3b8897288bb13bad4ca999fedfb28a3e Mon Sep 17 00:00:00 2001
From: Pepping <pepping>
Date: Mon, 26 Jan 2015 14:13:42 +0000
Subject: [PATCH] - Added c_force_late_sync and c_force_early_sync - Fixed
 wrong indent of ref_data_total

---
 .../reorder/tb/python/tc_reorder_transpose.py | 41 ++++++++++---------
 1 file changed, 21 insertions(+), 20 deletions(-)

diff --git a/libraries/base/reorder/tb/python/tc_reorder_transpose.py b/libraries/base/reorder/tb/python/tc_reorder_transpose.py
index d08397e391..d5ebd7e144 100644
--- a/libraries/base/reorder/tb/python/tc_reorder_transpose.py
+++ b/libraries/base/reorder/tb/python/tc_reorder_transpose.py
@@ -88,8 +88,11 @@ c_db_nof_streams    = c_bg_nof_streams
 c_db_ram_size       = c_bg_ram_size #g_rd_chunksize * g_rd_nof_chunks * g_nof_blocks
 c_frame_size        = g_wr_chunksize
 c_nof_int_streams   = 1   
-c_ena_pre_transpose = False
+c_ena_pre_transpose = True
 c_gap_size          = 0 #g_rd_chunksize
+c_force_late_sync   = 0
+c_force_early_sync  = 0
+
 
 tc.append_log(3, '>>>')
 tc.append_log(1, '>>> Title : Test bench for ddr3_transpose' )
@@ -121,7 +124,7 @@ def gen_bg_hex_files(c_nof_values = 1024, c_nof_streams = 4):
         stream_im = []
         for j in range(c_nof_values): 
             stream_re.append(j)
-            stream_im.append(i)
+            stream_im.append(i)  
         data_concat = dsp_test_bg.concatenate_two_lists(stream_re, stream_im, c_in_dat_w)
         data.append(data_concat)
         filename = "../../src/hex/tb_bg_dat_" + str(i) + ".hex"
@@ -139,11 +142,11 @@ if __name__ == "__main__":
         for j in range(g_rd_chunksize):
             ss_list.append(i + j*c_frame_size)
             
-
+    if c_ena_pre_transpose:
+        ss.write_selects(ss_list) 
+        
 #    for i in ss_list:
-#        print i
-#    ss.write_selects(ss_list)
-    
+#        print i           
     ###############################################################################
     #
     # Create stimuli for the BG
@@ -179,12 +182,17 @@ if __name__ == "__main__":
     do_until_gt(io.simIO.getSimTime, ms_retry=1000, val=110000, s_timeout=13600)  # 110000 
         
     # Enable the blockgenerator
-    bg.write_enable()
-    do_until_gt(io.simIO.getSimTime, ms_retry=1000, val=180000, s_timeout=13600)  # 110000 
-    bg.write_disable()
+    bg.write_enable()   
     
-    do_until_gt(io.simIO.getSimTime, ms_retry=1000, val=182000, s_timeout=13600)  # 110000 
-    bg.write_enable()
+    if(c_force_late_sync == 1): 
+        do_until_gt(io.simIO.getSimTime, ms_retry=1000, val=180000, s_timeout=13600)  # 110000 
+        bg.write_block_gen_settings(samplesPerPacket=c_frame_size, blocksPerSync=g_nof_blk_per_sync+1, gapSize=c_gap_size, memLowAddr=0, memHighAddr=c_bg_ram_size-1, BSNInit=10)
+    elif(c_force_early_sync == 1):
+        do_until_gt(io.simIO.getSimTime, ms_retry=1000, val=180000, s_timeout=13600)  # 110000 
+        bg.write_block_gen_settings(samplesPerPacket=c_frame_size, blocksPerSync=g_nof_blk_per_sync-1, gapSize=c_gap_size, memLowAddr=0, memHighAddr=c_bg_ram_size-1, BSNInit=10)
+    
+    #do_until_gt(io.simIO.getSimTime, ms_retry=1000, val=182000, s_timeout=13600)  # 110000 
+    #bg.write_enable()
     #
     #do_until_gt(io.simIO.getSimTime, ms_retry=1000, val=120000, s_timeout=13600)  # 110000 
     #bg.write_disable()
@@ -197,8 +205,6 @@ if __name__ == "__main__":
     #
     ###############################################################################
     # Subband Select pre-transpose 
-    print "len(ss_list)"            
-    print len(ss_list) 
     if c_ena_pre_transpose:
         bg_data = ss.subband_select(bg_data, ss_list)    
     
@@ -208,11 +214,6 @@ if __name__ == "__main__":
         bg_data_single_page = bg_data[t*c_pagesize:(t+1)*c_pagesize]
         # Write to memory
         mem_page = [0] * c_pagesize   
-        print "len(mem_page)"                
-        print len(mem_page)                
-        print "len(bg_data_single_page)"
-        print len(bg_data_single_page)                
-        
         for i in range(g_nof_blocks):
             for j in range(g_wr_nof_chunks):
                 for k in range(g_wr_chunksize):
@@ -231,10 +232,10 @@ if __name__ == "__main__":
                     #ref_data[chunk_cnt*(g_rd_chunksize*g_rd_nof_chunks)+ chunk_offset*g_rd_chunksize + k] = mem_page[chunk_cnt*(g_rd_chunksize*g_rd_nof_chunks+g_gapsize)+ chunk_offset*g_rd_chunksize + k]
                     ref_data[i*(g_rd_chunksize*g_rd_nof_chunks)+j*g_rd_chunksize + k] = mem_page[chunk_cnt*(g_rd_chunksize*g_rd_nof_chunks+g_gapsize)+ chunk_offset*g_rd_chunksize + k]
                 chunk_cnt = chunk_cnt + 1    
-            ref_data_total.append(ref_data)
+        ref_data_total.append(ref_data)
         
     ref_data_total=flatten(ref_data_total)    
-      
+
     # Split the data again in individual channels 
     ref_data_split = []
     ref_data_split.append(ref_data_total)
-- 
GitLab