Skip to content
GitLab
Explore
Sign in
Register
Primary navigation
Search or go to…
Project
H
HDL
Manage
Activity
Members
Labels
Plan
Issues
Issue boards
Milestones
Iterations
Wiki
Requirements
Jira
Code
Merge requests
Repository
Branches
Commits
Tags
Repository graph
Compare revisions
Snippets
Locked files
Build
Pipelines
Jobs
Pipeline schedules
Test cases
Artifacts
Deploy
Releases
Container registry
Model registry
Operate
Environments
Monitor
Incidents
Analyze
Value stream analytics
Contributor analytics
CI/CD analytics
Repository analytics
Code review analytics
Issue analytics
Insights
Model experiments
Help
Help
Support
GitLab documentation
Compare GitLab plans
GitLab community forum
Contribute to GitLab
Provide feedback
Keyboard shortcuts
?
Snippets
Groups
Projects
Show more breadcrumbs
RTSD
HDL
Commits
855c8f93
Commit
855c8f93
authored
6 years ago
by
Eric Kooistra
Browse files
Options
Downloads
Patches
Plain Diff
Tested script and used it to start WG on two dishes.
parent
150fbfdf
Branches
Branches containing commit
No related tags found
No related merge requests found
Changes
1
Show whitespace changes
Inline
Side-by-side
Showing
1 changed file
applications/apertif/commissioning/tests/start_wg_at_bsn.sh
+23
-8
23 additions, 8 deletions
applications/apertif/commissioning/tests/start_wg_at_bsn.sh
with
23 additions
and
8 deletions
applications/apertif/commissioning/tests/start_wg_at_bsn.sh
+
23
−
8
View file @
855c8f93
...
...
@@ -22,21 +22,36 @@
# Author: Eric Kooistra
# Purpose: Start WG synchronously at multiple dishes
# Usage:
# . First use util_bsn_source.py on a dish to read the current BSN
#
#
ccu-corr
> ssh lcu-rtc -X "python $UPE/peripherals/util_bsn_source.py --unb 0 --bn 0 -n 7" # read current BSN @ sync
#
At
ccu-corr
.
#
# . The BSN increments by 800000 every 1.024 sec, so e.g. start WG about 30 sec from now:
# start BSN = current BSN + 30 * 800000
# . Then start the WG at all dishes using:
# Edit this script to set the desired WG amplitude and frequency via subband, channel
#
#
ccu-corr> python tests/run_command_on_lcus.py --tel 2,3 -s 'start_wg_at_bsn.sh <start BSN>'
#
First use util_bsn_source.py on a dish to read the <current BSN> at sync
#
# ssh lcu-rtc -X "python $UPE/peripherals/util_bsn_source.py --unb 0 --bn 0 -n 7"
#
# The BSN increments by 800000 every 1.024 sec, so e.g. start WG 30 sync periods of 1.024s from now:
#
# Calcuclate <start BSN> = <current BSN> + 30 * 800000
#
# To start the WG on a single dish use:
#
# ssh lcu-rtc -X "start_wg_at_bsn.sh <start BSN>"
#
# To start the WG at all dishes use:
#
# python tests/run_command_on_lcus.py --tel 2,3 -s '$RADIOHDL/applications/apertif/commissioning/tests/start_wg_at_bsn.sh <start BSN>'
#
# Read SST to check that WG is active
#
# ssh lcu-rtc -X "python $UPE/peripherals/pi_apertif_system.py --unb 0:7 --bn 0:3 --sp 0:3 --cmd 10 -v 5 --useplot"
# User parameters
SCHEDULED_BSN
=
$1
WG_AMPL
=
10
WG_SUB
=
64
# subband 64 + n*24 arrives at central UniBoard n
WG_CHAN
=
32
# channel in subband 0..63, where 32 is at center
WG_CHAN_FRACTION
=
0.0
# channel fraction 0.0 is at center of channel
...
...
@@ -56,7 +71,7 @@ fi
python
$UPE
/peripherals/pi_apertif_system.py
--unb
0:7
--bn
0:3
--sp
0:3
--pol
0,1
--cmd
20
-v
5
--globalsp
0:63
--sub
0
--chan
32
--ampl
0
--phs
0
# Setup WG_AMPL for one GLOBAL_SP and one POL
python
$UPE
/peripherals/pi_apertif_system.py
--unb
0:3
--bn
0:3
--sp
0:3
--pol
$POL
--cmd
20
-v
5
--globalsp
$GLOBAL_SP
--sub
$WG_SUB
--chan
WG_CHAN
--ampl
WG_AMPL
--chanfraction
$WG_CHAN_FRACTION
--phs
WG_PHASE
python
$UPE
/peripherals/pi_apertif_system.py
--unb
0:3
--bn
0:3
--sp
0:3
--pol
$POL
--cmd
20
-v
5
--globalsp
$GLOBAL_SP
--sub
$WG_SUB
--chan
$
WG_CHAN
--ampl
$
WG_AMPL
--chanfraction
$WG_CHAN_FRACTION
--phs
$
WG_PHASE
# Start WG at scheduled BSN in future
python
$UPE
/peripherals/util_diag_wg_wideband.py
--unb
0:7
--bn
0:3
--sp
0:3
-v
3
-n
1
--bsn
$SCHEDULED_BSN
python
$UPE
/peripherals/util_diag_wg_wideband.py
--unb
0:7
--bn
0:3
--sp
0:3
-v
5
-n
1
--bsn
$SCHEDULED_BSN
This diff is collapsed.
Click to expand it.
Preview
0%
Loading
Try again
or
attach a new file
.
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Save comment
Cancel
Please
register
or
sign in
to comment