diff --git a/notebooks_tsqr/NightLog.ipynb b/notebooks_tsqr/NightLog.ipynb index 98e2054..298c54f 100644 --- a/notebooks_tsqr/NightLog.ipynb +++ b/notebooks_tsqr/NightLog.ipynb @@ -10,7 +10,7 @@ }, { "cell_type": "code", - "execution_count": 1, + "execution_count": null, "id": "1", "metadata": {}, "outputs": [], @@ -33,7 +33,7 @@ }, { "cell_type": "code", - "execution_count": 2, + "execution_count": null, "id": "2", "metadata": { "jupyter": { @@ -78,7 +78,7 @@ }, { "cell_type": "code", - "execution_count": 3, + "execution_count": null, "id": "3", "metadata": {}, "outputs": [], @@ -95,7 +95,7 @@ }, { "cell_type": "code", - "execution_count": 4, + "execution_count": null, "id": "4", "metadata": { "jupyter": { @@ -148,43 +148,10 @@ }, { "cell_type": "code", - "execution_count": 5, + "execution_count": null, "id": "7", "metadata": {}, - "outputs": [ - { - "data": { - "text/markdown": [ - "\n", - "Report for **2024-09-18** covering the previous **3** observing night(s).\n", - "- Run on logs from **https://summit-lsp.lsst.codes/**\n", - "- Using *Prototype* Logging and Reporting Version: **0.1.dev63+g83a61e0.d20240918**\n" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "This report will attempt to use the following log sources: \n", - "- https://summit-lsp.lsst.codes/exposurelog/instruments\n", - "- https://summit-lsp.lsst.codes/exposurelog/exposures\n", - "- https://summit-lsp.lsst.codes/exposurelog/messages\n", - "- https://summit-lsp.lsst.codes/narrativelog/messages\n", - "- https://summit-lsp.lsst.codes/nightreport/reports" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - } - ], + "outputs": [], "source": [ "# Display overview of Report context \n", "md(f'''\n", @@ -199,18 +166,10 @@ }, { "cell_type": "code", - "execution_count": 6, + "execution_count": null, "id": "8", "metadata": {}, - "outputs": [ - { - "name": "stdout", - "output_type": "stream", - "text": [ - "\n" - ] - } - ], + "outputs": [], "source": [ "print()" ] @@ -225,119 +184,14 @@ }, { "cell_type": "code", - "execution_count": 7, + "execution_count": null, "id": "10", "metadata": { "jupyter": { "source_hidden": true } }, - "outputs": [ - { - "data": { - "text/html": [ - "
\n", - "\n", - "\n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - " \n", - "
01
Moon Rise2024-09-17 22:25:43.601
Moon Set2024-09-17 10:17:44.848
Moon Illumination99%(% lit)
Astronomical Twilight (morning)2024-09-17 09:18:25.687(-18 degrees)
Astronomical Twilight (evening)2024-09-17 23:56:37.079(-18 degrees)
Nautical Twilight (morning)2024-09-17 09:46:22.573(-12 degrees)
Nautical Twilight (evening)2024-09-17 23:28:37.050(-12 degrees)
Civil Twilight (morning)2024-09-17 10:14:10.623(-6 degrees)
Civil Twilight (evening)2024-09-17 23:00:46.504(-6 degrees)
Sun Rise2024-09-17 10:41:56.478
Sun Set2024-09-17 22:32:58.645
\n", - "
" - ], - "text/plain": [ - " 0 1\n", - "Moon Rise 2024-09-17 22:25:43.601 \n", - "Moon Set 2024-09-17 10:17:44.848 \n", - "Moon Illumination 99% (% lit)\n", - "Astronomical Twilight (morning) 2024-09-17 09:18:25.687 (-18 degrees)\n", - "Astronomical Twilight (evening) 2024-09-17 23:56:37.079 (-18 degrees)\n", - "Nautical Twilight (morning) 2024-09-17 09:46:22.573 (-12 degrees)\n", - "Nautical Twilight (evening) 2024-09-17 23:28:37.050 (-12 degrees)\n", - "Civil Twilight (morning) 2024-09-17 10:14:10.623 (-6 degrees)\n", - "Civil Twilight (evening) 2024-09-17 23:00:46.504 (-6 degrees)\n", - "Sun Rise 2024-09-17 10:41:56.478 \n", - "Sun Set 2024-09-17 22:32:58.645 " - ] - }, - "execution_count": 7, - "metadata": {}, - "output_type": "execute_result" - } - ], + "outputs": [], "source": [ "# Display various almanac values (for moon, sun)\n", "rep.AlmanacReport().almanac_as_dataframe()" @@ -353,947 +207,14 @@ }, { "cell_type": "code", - "execution_count": 8, + "execution_count": null, "id": "12", "metadata": { "jupyter": { "source_hidden": true } }, - "outputs": [ - { - "data": { - "text/markdown": [ - "## Nightly Jira BLOCKs\n", - "- 20240916\n", - " - [testPlayer/BLOCK-R89](https://rubinobs.atlassian.net/projects/BLOCK?selectedItem=com.atlassian.plugins.atlassian-connect-plugin:com.kanoah.test-manager__main-project-page#!/testPlayer/BLOCK-R89)\n", - " - [v2/testCases](https://rubinobs.atlassian.net/projects/BLOCK?selectedItem=com.atlassian.plugins.atlassian-connect-plugin:com.kanoah.test-manager__main-project-page#!/v2/testCases)\n", - "- 20240917\n", - " - [v2/testCase/BLOCK-T19](https://rubinobs.atlassian.net/projects/BLOCK?selectedItem=com.atlassian.plugins.atlassian-connect-plugin:com.kanoah.test-manager__main-project-page#!/v2/testCase/BLOCK-T19)\n", - " - [testPlayer/BLOCK-R90](https://rubinobs.atlassian.net/projects/BLOCK?selectedItem=com.atlassian.plugins.atlassian-connect-plugin:com.kanoah.test-manager__main-project-page#!/testPlayer/BLOCK-R90)" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "### " - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "## NIGHT: 20240916: " - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "### DATE: 2024-09-16: " - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:18:03 |
Daytime: Cycle 39 started around 09:00 CLT, so most of CSCs were in Offline or Standby until it was done. \n",
-       "\n",
-       "We run successfully all the blocks. \n",
-       "- BLOCK-T151 (1.0) Simonyi Hexapods Startup. Both M2/Camera Hexapods shows 17.3 C.\n",
-       "- BLOCK-T3 (2.0) Warm-up M2 Hexapod.\n",
-       "- BLOCK-T4 (2.0) Warm-up Camera Hexapod.\n",
-       "- BLOCK-T153 - Opening M1M3 Mirror Covers.\n",
-       "- BLOCK-T152  - Closing M1M3 Mirror Covers.\n",
-       "- BLOCK-T145 - M1M3 Hardpoints Breakaway Test.\n",
-       "- BLOCK-T144 - M1M3 Bump Test - Actuators [238, 330, 409] FAILED the bump test.\n",
-       "- BLOCK-T139 - Hexapod Shutdown.\n",
-       "\n",
-       "The ScriptQueue:1 was updated and patched during the night. After the warm-up, M2 and Camera Hexapod were in disabled state during all the run and we monitor their temperature. Since the end of the warm-up until the hexapod shutdown their temperature increased ~ 1.5 deg C (average).
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:18:03 |
Daytime: Cycle 39 started around 09:00 CLT, so most of CSCs were in Offline or Standby until it was done. \n",
-       "\n",
-       "We run successfully all the blocks. \n",
-       "- BLOCK-T151 (1.0) Simonyi Hexapods Startup. Both M2/Camera Hexapods shows 17.3 C\n",
-       "- BLOCK-T3 (2.0) Warm-up M2 Hexapod\n",
-       "- BLOCK-T4 (2.0) Warm-up Camera Hexapod\n",
-       "- We set DISABLED state for both hexapod and monitor their temperature (temperature increased ~ 1.5 deg C from 18 to 19.4)\n",
-       "- BLOCK-T153 - Opening M1M3 Mirror Covers\n",
-       "- BLOCK-T152  - Closing M1M3 Mirror Covers \n",
-       "- BLOCK-T145 - M1M3 Hardpoints Breakaway Test\n",
-       "- Tiago updated ScriptQueue:1\n",
-       "- BLOCK-T144 - M1M3 Bump Test - Actuators [238, 330, 409] FAILED the bump test\n",
-       "- BLOCK-T139 - Hexapod Shutdown
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:18:03 |
Daytime: Cycle 39 started around 09:00 CLT, so most of CSCs were in Offline or Standby until it was done. \n",
-       "\n",
-       "We run successfully all the blocks. \n",
-       "- BLOCK-T151 (1.0) Simonyi Hexapods Startup. Both M2/Camera Hexapods shows 17.3 C.\n",
-       "- BLOCK-T3 (2.0) Warm-up M2 Hexapod.\n",
-       "- BLOCK-T4 (2.0) Warm-up Camera Hexapod.\n",
-       "- BLOCK-T153 - Opening M1M3 Mirror Covers.\n",
-       "- BLOCK-T152  - Closing M1M3 Mirror Covers.\n",
-       "- BLOCK-T145 - M1M3 Hardpoints Breakaway Test.\n",
-       "- BLOCK-T144 - M1M3 Bump Test - Actuators [238, 330, 409] FAILED the bump test.\n",
-       "- BLOCK-T139 - Hexapod Shutdown.\n",
-       "\n",
-       "The ScriptQueue:1 was updated and patched during the night. After the warm-up, M2 and Camera Hexapod were in disabled state during all the run and we monitor their temperature. Since the end of the warm-up until the hexapod shutdown their temperature increased ~ 1.5 deg C (average).
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:18:03 |
Daytime: Cycle 39 started around 09:00 CLT, so most of CSCs were in Offline or Standby until it was done. Tony J. restart the ATcamera processes to prepare for cycle 39 upgrade. \n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:18:03 |
Daytime: Cycle 39 started around 09:00 CLT, so most of CSCs were in Offline or Standby until it was done. Tony J. restart the auxtel camera processes to prepare for cycle 39 upgrade. \n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:18:03 |
Daytime: Cycle 39 started around 09:00 CLT, so most of CSCs were in Offline or Standby until it was done. \n",
-       "\n",
-       "We run successfully all the blocks. \n",
-       "- BLOCK-T151 (1.0) Simonyi Hexapods Startup. Both M2/Camera Hexapods shows 17.3 C\n",
-       "- BLOCK-T3 (2.0) Warm-up M2 Hexapod\n",
-       "- BLOCK-T4 (2.0) Warm-up Camera Hexapod\n",
-       "- We set DISABLED state for both hexapod and monitor their temperature (temperature increased ~ 1.5 deg C from 18 to 19.4)\n",
-       "- BLOCK-T153 - Opening M1M3 Mirror Covers\n",
-       "- BLOCK-T152  - Closing M1M3 Mirror Covers \n",
-       "- BLOCK-T145 - M1M3 Hardpoints Breakaway Test\n",
-       "- Tiago updated ScriptQueue:1\n",
-       "- BLOCK-T144 - M1M3 Bump Test - Actuators [238, 330, 409] FAILED the bump test\n",
-       "- BLOCK-T139 - Hexapod Shutdown
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:18:03 |
Daytime: Cycle 39 started around 09:00 CLT, so most of CSCs were in Offline or Standby until it was done. \n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:18:03 |
Daytime: Cycle 39 started around 09:00 CLT, so most of CSCs were in Offline or Standby until it was done. \n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 13:44:14 |
Daytime: Cycle 39 started around 09:00 CLT, so most of CSCs were in Offline or Standby until it was done. Tony J. restart the auxtel camera processes to prepare for cycle 39 upgrade. \n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 13:44:14 |
Daytime: Cycle 39 started around 09:00 CLT, so most CSCs were Offline or Standby until it was done. Tony J. restarted the Auxtel camera processes to prepare for the cycle 39 upgrade. At the beginning of checkout, we had an image Timeout waiting for imageInOODS event issue with latiss_checkout. Steve P. helped figure it out and found that there was an issue with /data/staging/auxtel/oods at nfs-auxtel.cp.lsst.org, which was responding very slowly at that time. As this issue keeps happening (OBS-343), we shared this issue in #rubinobs-it-chile channel. \n",
-       "\n",
-       "While running power_on_atcalsys.py, we had an issue with one of the limit switches on the ATWhitelight (\"Shutter failed to open\" message on CSC) and it was solved when we manually moved the switch and made it touch the shutter. Latiss daytime calibration was done even though it failed at some attempts due to missing a patch for external scripts.  \n",
-       "\n",
-       "Around 20:40 pm CLT, AuxTel was ready to be on Sky and start the scheduler at 21 pm CLT after ATScriptQueue was updated. \n",
-       "\n",
-       "There is one created ticket to report an issue with linking the JIRA ticket to OLE (LOVE-362)
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 13:44:14 |
Daytime: Cycle 39 started around 09:00 CLT, so most of the CSCs were in Offline or Standby until it was done. Tony J. restarted the auxtel camera processes to prepare for the cycle 39 upgrade. At the beginning of checkout, we had an image Timeout waiting for imageInOODS event issue with latiss_checkout. Steve P. helped figure it out and found that there was an issue with /data/staging/auxtel/oods at nfs-auxtel.cp.lsst.org, which was responding very slowly at that time. As this issue keeps happening (OBS-343), we shared this issue in #rubinobs-it-chile channel. \n",
-       "\n",
-       "While running power_on_atcalsys.py, we had an issue with \n",
-       "\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 13:44:14 |
Daytime: Cycle 39 started around 09:00 CLT, so most of the CSCs were in Offline or Standby until it was done. Tony J. restarted the Auxtel camera processes to prepare for the cycle 39 upgrade. At the beginning of checkout, we had an image Timeout waiting for imageInOODS event issue with latiss_checkout. Steve P. helped figure it out and found that there was an issue with /data/staging/auxtel/oods at nfs-auxtel.cp.lsst.org, which was responding very slowly at that time. As this issue keeps happening (OBS-343), we shared this issue in #rubinobs-it-chile channel. \n",
-       "\n",
-       "While running power_on_atcalsys.py, we had an issue with one of the limit switches on the ATWhitelight (\"Shutter failed to open\" message on CSC) and it was solved when we manually moved the switch and made it touch the shutter. Latiss daytime calibration was done even though it failed at the beginning due to missing a patch for external scripts.  \n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 13:44:14 |
Daytime: Cycle 39 started around 09:00 CLT, so most CSCs were Offline or Standby until it was done. Tony J. restarted the Auxtel camera processes to prepare for the cycle 39 upgrade. At the beginning of checkout, we had an image Timeout waiting for imageInOODS event issue with latiss_checkout. Steve P. helped figure it out and found that there was an issue with /data/staging/auxtel/oods at nfs-auxtel.cp.lsst.org, which was responding very slowly at that time. As this issue keeps happening (OBS-343), we shared this issue in #rubinobs-it-chile channel. \n",
-       "\n",
-       "While running power_on_atcalsys.py, we had an issue with one of the limit switches on the ATWhitelight (\"Shutter failed to open\" message on CSC) and it was solved when we manually moved the switch and made it touch the shutter. Latiss daytime calibration was done even though it failed at some attempts due to missing a patch for external scripts.  \n",
-       "\n",
-       "Around 20:40 pm CLT, AuxTel was ready to be on sky and start the scheduler at pm CLT after ATScriptQueue updated. \n",
-       "\n",
-       "There is one created ticket to report an issue with linking JIRA ticket to OLE (LOVE-362)
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 13:44:14 |
Daytime: Cycle 39 started around 09:00 CLT, so most CSCs were Offline or Standby until it was done. Tony J. restarted the Auxtel camera processes to prepare for the cycle 39 upgrade. At the beginning of checkout, we had an image Timeout waiting for imageInOODS event issue with latiss_checkout. Steve P. helped figure it out and found that there was an issue with /data/staging/auxtel/oods at nfs-auxtel.cp.lsst.org, which was responding very slowly at that time. As this issue keeps happening (OBS-343), we shared this issue in #rubinobs-it-chile channel. \n",
-       "\n",
-       "While running power_on_atcalsys.py, we had an issue with one of the limit switches on the ATWhitelight (\"Shutter failed to open\" message on CSC) and it was solved when we manually moved the switch and made it touch the shutter. LATISS daytime calibrations were completed even though it failed at some attempts due to missing a patch for external scripts.  \n",
-       "\n",
-       "Around 20:40 pm CLT, AuxTel was ready to be on Sky and started the scheduler at 21 pm CLT after ATScriptQueue was updated and patched. \n",
-       "\n",
-       "There is one created ticket to report an issue on linking a JIRA ticket directly from OLE (LOVE-362).\n",
-       "\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 13:44:14 |
Daytime: Cycle 39 started around 09:00 CLT, so most CSCs were Offline or Standby until it was done. Tony J. restarted the Auxtel camera processes to prepare for the cycle 39 upgrade. At the beginning of checkout, we had an image Timeout waiting for imageInOODS event issue with latiss_checkout. Steve P. helped figure it out and found that there was an issue with /data/staging/auxtel/oods at nfs-auxtel.cp.lsst.org, which was responding very slowly at that time. As this issue keeps happening (OBS-343), we shared this issue in #rubinobs-it-chile channel. \n",
-       "\n",
-       "While running power_on_atcalsys.py, we had an issue with one of the limit switches on the ATWhitelight (\"Shutter failed to open\" message on CSC) and it was solved when we manually moved the switch and made it touch the shutter. LATISS daytime calibrations were completed even though it failed at some attempts due to missing a patch for external scripts.  \n",
-       "\n",
-       "Around 20:40 pm CLT, AuxTel was ready to be on Sky and started the scheduler at 21 pm CLT after ATScriptQueue was updated and patched. \n",
-       "\n",
-       "There is one created ticket to report an issue on linking a JIRA ticket directly from OLE (LOVE-362).\n",
-       "\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 13:44:14 |
Daytime: Cycle 39 started around 09:00 CLT, so most CSCs were Offline or Standby until it was done. Tony J. restarted the Auxtel camera processes to prepare for the cycle 39 upgrade. At the beginning of checkout, we had an image Timeout waiting for imageInOODS event issue with latiss_checkout. Steve P. helped figure it out and found that there was an issue with /data/staging/auxtel/oods at nfs-auxtel.cp.lsst.org, which was responding very slowly at that time. As this issue keeps happening (OBS-343), we shared this issue in #rubinobs-it-chile channel. \n",
-       "\n",
-       "While running power_on_atcalsys.py, we had an issue with one of the limit switches on the ATWhitelight (\"Shutter failed to open\" message on CSC) and it was solved when we manually moved the switch and made it touch the shutter. Latiss daytime calibration was done even though it failed at some attempts due to missing a patch for external scripts.  \n",
-       "\n",
-       "Around 20:40 pm CLT, AuxTel was ready to be on Sky and start the scheduler at 21 pm CLT after ATScriptQueue was updated. \n",
-       "\n",
-       "There is one created ticket to report an issue with linking the JIRA ticket to OLE (LOVE-362)
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 13:44:14 |
Daytime: Cycle 39 started around 09:00 CLT, so most CSCs were Offline or Standby until it was done. Tony J. restarted the Auxtel camera processes to prepare for the cycle 39 upgrade. At the beginning of checkout, we had an image Timeout waiting for imageInOODS event issue with latiss_checkout. Steve P. helped figure it out and found that there was an issue with /data/staging/auxtel/oods at nfs-auxtel.cp.lsst.org, which was responding very slowly at that time. As this issue keeps happening (OBS-343), we shared this issue in #rubinobs-it-chile channel. \n",
-       "\n",
-       "While running power_on_atcalsys.py, we had an issue with one of the limit switches on the ATWhitelight (\"Shutter failed to open\" message on CSC) and it was solved when we manually moved the switch and made it touch the shutter. Latiss daytime calibration was done even though it failed at some attempts due to missing a patch for external scripts.  \n",
-       "\n",
-       "\n",
-       "\n",
-       "\n",
-       "There is one created ticket to report an issue with linking JIRA ticket to OLE (LOVE-362)
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 13:44:14 |
Daytime: Cycle 39 started around 09:00 CLT, so most of the CSCs were in Offline or Standby until it was done. Tony J. restarted the Auxtel camera processes to prepare for the cycle 39 upgrade. At the beginning of checkout, we had an image Timeout waiting for imageInOODS event issue with latiss_checkout. Steve P. helped figure it out and found that there was an issue with /data/staging/auxtel/oods at nfs-auxtel.cp.lsst.org, which was responding very slowly at that time. As this issue keeps happening (OBS-343), we shared this issue in #rubinobs-it-chile channel. \n",
-       "\n",
-       "While running power_on_atcalsys.py, we had an issue with one of the limit switches on the ATWhitelight (\"Shutter failed to open\" message on CSC) and it was solved when we manually moved the switch and made it touch the shutter. Latiss daytime calibration was done even though it failed at the beginning due to missing a patch for external scripts.  \n",
-       "\n",
-       "\n",
-       "\n",
-       "There is one created ticket to report an issue with linking JIRA ticket to OLE (LOVE-362)
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 13:44:14 |
Daytime: Cycle 39 started around 09:00 CLT, so most of the CSCs were in Offline or Standby until it was done. Tony J. restarted the auxtel camera processes to prepare for the cycle 39 upgrade. At the beginning of checkout, we had an image Timeout waiting for imageInOODS event issue with latiss_checkout. Steve P. helped figure it out and found that there was an issue with /data/staging/auxtel/oods at nfs-auxtel.cp.lsst.org, which was responding very slowly at that time. As this issue keeps happening (OBS-343), we shared this issue in #rubinobs-it-chile channel. \n",
-       "\n",
-       "While running power_on_atcalsys.py, we had an issue with one of the limit switches on the ATWhitelight (\"Shutter failed to open\" message on CSC) and it was solved when we manually moved the switch and made it touch the shutter. Latiss daytime calibration was done even though it failed at the beginning due to missing a patch for external scripts.  \n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 13:44:14 |
Daytime: Cycle 39 started around 09:00 CLT, so most of CSCs were in Offline or Standby until it was done. Tony J. restart the auxtel camera processes to prepare for cycle 39 upgrade. \n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "## NIGHT: 20240917: " - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "### DATE: 2024-09-17: " - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:45:42 |
Daytime: In the morning Craig L. checked the limit switch (open) of the ATWhiteLight shutter and adjusted it. Now the limit switch works properly. Around 11:00 am CLT, Daytime checkout proceeded and was done without issue. Venting started around 11:30 am CLT. Around 15 pm CLT, we stopped venting and started calibration. Around 16:15pm CLT, we finished calibration and restarted venting until the beginning of the night. \n",
-       "\n",
-       "\n",
-       "Nighttime: When preparing for on sky scheduler, Scheduler:2 went to FAULT state. This is currently a bug in the scheduler:2. The solution is to wait until there are targets available. \n",
-       "\n",
-       "We open the dome later (01:00 UT) because it was overcast.
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:45:42 |
Daytime: In the morning Craig L. checked the limit switch (open) of the ATWhiteLight shutter and adjusted it. Now the limit switch works properly. Around 11:00 am CLT, Daytime checkout proceeded and was done without issue. Venting started around 11:30 am CLT. Around 15 pm CLT, we stopped venting and started calibration. Around 16:15pm CLT, we finished calibration and restarted venting until the beginning of the night. \n",
-       "\n",
-       "\n",
-       "Nighttime: When preparing for on sky scheduler, Scheduler:2 went to FAULT state. There was a bug identified on the Scheduler:2 and now it is solved. On-sky observations started around 22:00 CLT when only thin clouds remained at the horizon.\n",
-       "\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:45:42 |
Daytime: In the morning Craig L. checked the limit switch (open) of the ATWhiteLight shutter and adjusted it. Now the limit switch works properly. Around 11:00 am CLT, Daytime checkout proceeded and was done without issue. Venting started around 11:30 am CLT. Around 15 pm CLT, we stopped venting and started calibration. Around 16:15pm CLT, we finished calibration and restarted venting until the beginning of the night. \n",
-       "\n",
-       "\n",
-       "Nighttime: When preparing for on sky scheduler, Scheduler:2 went to FAULT state. This is currently a bug in the scheduler:2. The solution is to wait until there are targets available. We had to close the dome as it was overcast.
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:45:42 |
Daytime: In the morning Craig L. checked the limit switch (open) of the ATWhiteLight shutter and adjusted it. Now the limit switch works properly. Around 11:00 am CLT, Daytime checkout proceeded and was done without issue. Venting started around 11:30 am CLT. Around 15 pm CLT, we stopped venting and started calibration. Around 16:15pm CLT, we finished calibration and restarted venting until the beginning of the night. \n",
-       "\n",
-       "\n",
-       "Nighttime: Scheduler:2 went to fault 
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:45:42 |
Daytime: In the morning Craig L. checked the limit switch (open) of the ATWhiteLight shutter and adjusted it. Now the limit switch works properly. Around 11:00am CLT, Daytime checkout proceeded and done without issue. Venting started around 11:30 am CLT. \n",
-       "\n",
-       "\n",
-       "\n",
-       "\n",
-       "Nighttime:
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:45:42 |
Daytime: In the morning Craig L. checked the limit switch (open) of the ATWhiteLight shutter and adjusted it. Now the limit switch works properly. Around 11:00 am CLT, Daytime checkout proceeded and was done without issue. Venting started around 11:30 am CLT. Around 15 pm CLT, we stopped venting and started calibration. \n",
-       "\n",
-       "\n",
-       "\n",
-       "Nighttime:
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:45:42 |
Daytime:\n",
-       "\n",
-       "\n",
-       "\n",
-       "\n",
-       "Nighttime:
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:45:42 |
Daytime: In the morning Craig L. checked the limit switch (open) of the ATWhiteLight shutter and adjusted it. Now the limit switch works properly. Around 11:00 am CLT, Daytime checkout proceeded and was done without issue. Venting started around 11:30 am CLT. Around 15 pm CLT, we stopped venting and started calibration. Around 16:15pm CLT, we finished calibration and restarted venting until the beginning of the night. \n",
-       "\n",
-       "\n",
-       "Nighttime: When preparing for on sky scheduler, Scheduler:2 went to FAULT state. There was a bug identified on the Scheduler:2 and now it is solved. On-sky observations started around 22:00 CLT when only thin clouds remained at the horizon.\n",
-       "\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:45:42 |
Daytime: In the morning Craig L. checked the limit switch (open) of the ATWhiteLight shutter and adjusted it. Now the limit switch works properly. Around 11:00 am CLT, Daytime checkout proceeded and was done without issue. Venting started around 11:30 am CLT. Around 15 pm CLT, we stopped venting and started calibration. Around 16:15pm CLT, we finished calibration and restarted venting until the beginning of the night. \n",
-       "\n",
-       "\n",
-       "Nighttime:
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:45:42 |
"
-      ],
-      "text/plain": [
-       ""
-      ]
-     },
-     "metadata": {},
-     "output_type": "display_data"
-    },
-    {
-     "data": {
-      "text/markdown": [
-       "> 12:45:42 | 
Daytime: In the morning Craig L. checked the limit switch (open) of the ATWhiteLight shutter and adjusted it. Now the limit switch works properly. Around 11:00 am CLT, Daytime checkout proceeded and was done without issue. Venting started around 11:30 am CLT. Around 15 pm CLT, we stopped venting and started calibration. \n",
-       "\n",
-       "\n",
-       "\n",
-       "Nighttime:
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:46:16 |
During daytime, Petr K. worked on M1M3 to run bump tests, raise the mirror, and test to find the reference/center position for static support centering (see more details and results at #m1m3_worklog channel). \n",
-       "\n",
-       "We started from Simonyi Hexapods Startup (BLOCK-T151) around 18:30 UTC and it was done without any issues. Then warm-up M2 Hexapod (BLOCK-T3) and warm Camera Hexapod (BLOCK-T4) were completed successfully. Both components were left at enabled state and were monitored. Strut12 from Camera Hexapod showed a notorious increase on temperature (about 2C in about 20 min plus diverting in motor current from the rest of the struts). It was turned off and left in Standby state.\n",
-       "\n",
-       "\n",
-       "\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:46:16 |
During daytime, Petr K. worked on M1M3 to run bump tests, raise the mirror, and test to find the reference/center position for static support centering (see more details and results at #m1m3_worklog channel). \n",
-       "\n",
-       "We started from Simonyi Hexapods Startup (BLOCK-T151) around 18:30 UTC and it was done without any issues. Then warm-up M2 Hexapod (BLOCK-T3) and warm Camera Hexapod (BLOCK-T4) were completed successfully. Both components were left at enabled state and were monitored.\n",
-       "\n",
-       "Camera Hexapod Strut12 showed a notorious increase on temperature (about 2C in about 20 min plus diverting in motor current from the rest of the struts). It was turned off and left in Standby state. \n",
-       "\n",
-       "M2 Hexapod, we kept monitoring its currents (as temperature sensors is not available).\n",
-       "\n",
-       "Actuators [218, 238, 409] FAILED the bump test. \n",
-       "\n",
-       "\n",
-       "\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:46:16 |
During daytime, Petr K. worked on M1M3 to run bump tests, raise the mirror, and test to find the reference/center position for static support centering (see more details and results at #m1m3_worklog channel). \n",
-       "\n",
-       "We started the Simonyi Hexapods Startup (BLOCK-T151) around 18:30 UTC and it was done without any issues. Then warm-up M2 Hexapod (BLOCK-T3) and warm Camera Hexapod (BLOCK-T4) were completed successfully. Both components were left at enabled state and were monitored (see OBS-592).\n",
-       "\n",
-       "Camera Hexapod Strut12 showed a notorious increase on temperature (about 2C in about 20 min plus diverting in motor current from the rest of the struts). It was turned off and left in standby state. \n",
-       "\n",
-       "For M2 Hexapod, we kept monitoring its currents on enabled state (as temperature sensors are not available, OBS-593). At about 23:30 CLT, motor current on leg 4 approached the 4 amp. We proceed to shutdown the hexapod but it was not possible (either locally nor remote) to access the PDU with the known credentials (IHS-8426).\n",
-       "\n",
-       "The remaining Simonyi tests (BLOCK-T145 M1M3 Hardpoints Breakaway Test and BLOCK-T144 M1M3 Bump Test) run successfully. Actuators [218, 238, 409] FAILED the bump test. \n",
-       "\n",
-       "\n",
-       "\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:46:16 |
During daytime, Petr K. worked on M1M3 to run bump tests, raise the mirror, and test to find the reference/center position for static support centering (see more details and results at #m1m3_worklog channel). \n",
-       "\n",
-       "We started from Simonyi Hexapods Startup (BLOCK-T151) around 18:30 UTC and it was done without any issues. Then warm-up M2 Hexapod (BLOCK-T3) and warm Camera Hexapod (BLOCK-T4) were completed successfully. Both components were left at enabled state and were monitored.\n",
-       "\n",
-       "Camera Hexapod Strut12 showed a notorious increase on temperature (about 2C in about 20 min plus diverting in motor current from the rest of the struts). It was turned off and left in Standby state. \n",
-       "\n",
-       "M2 Hexapod, we kept monitoring its currents (as temperature sensors are not available).\n",
-       "\n",
-       "BLOCK-T145 M1M3 Hardpoints Breakaway Test run successfully.\n",
-       "\n",
-       "BLOCK-T144 M1M3 Bump Test, Actuators [218, 238, 409] FAILED the bump test. \n",
-       "\n",
-       "\n",
-       "\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:46:16 |
During daytime, Petr K. worked on M1M3 to run bump tests, raise the mirror, and test to find the reference/center position for static support centering (see more details and results at #m1m3_worklog channel). \n",
-       "\n",
-       "We started the Simonyi Hexapods Startup (BLOCK-T151) around 18:30 UTC and it was done without any issues. Then warm-up M2 Hexapod (BLOCK-T3) and warm Camera Hexapod (BLOCK-T4) were completed successfully. Both components were left at enabled state and were monitored (see OBS-592).\n",
-       "\n",
-       "Camera Hexapod Strut12 showed a notorious increase on temperature (about 2C in about 20 min plus diverting in motor current from the rest of the struts). It was turned off and left in standby state. \n",
-       "\n",
-       "For M2 Hexapod, we kept monitoring its currents on enabled state (as temperature sensors are not available, OBS-593). At about 23:30 CLT, motor current on leg 4 approached the 4 amp. We proceed to shutdown the hexapod but it was not possible (either locally nor remote) to access the PDU with the known credentials (IHS-8426).\n",
-       "\n",
-       "The remaining Simonyi tests (BLOCK-T145 M1M3 Hardpoints Breakaway Test and BLOCK-T144 M1M3 Bump Test) run successfully. Actuators [218, 238, 409] FAILED the bump test. \n",
-       "\n",
-       "\n",
-       "\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:46:16 |
During daytime, Petr K. worked on M1M3 to run bump tests, raise the mirror, and test to find the reference/center position for static support centering (see more details and results at #m1m3_worklog channel). \n",
-       "\n",
-       "We started the Simonyi Hexapods Startup (BLOCK-T151) around 18:30 UTC and it was done without any issues. Then warm-up M2 Hexapod (BLOCK-T3) and warm Camera Hexapod (BLOCK-T4) were completed successfully. Both components were left at enabled state and were monitored.\n",
-       "\n",
-       "Camera Hexapod Strut12 showed a notorious increase on temperature (about 2C in about 20 min plus diverting in motor current from the rest of the struts). It was turned off and left in standby state. \n",
-       "\n",
-       "M2 Hexapod, we kept monitoring its currents (as temperature sensors are not available).\n",
-       "\n",
-       "The remaining Simonyi tests (BLOCK-T145 M1M3 Hardpoints Breakaway Test and BLOCK-T144 M1M3 Bump Test) run successfully. Actuators [218, 238, 409] FAILED the bump test. \n",
-       "\n",
-       "\n",
-       "\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:46:16 |
During daytime, Petr K. worked on M1M3 to run bump tests, raise the mirror, and test to find the reference/center position for static support centering (see more details and results at #m1m3_worklog channel). \n",
-       "\n",
-       "We started the Simonyi Hexapods Startup (BLOCK-T151) around 18:30 UTC and it was done without any issues. Then warm-up M2 Hexapod (BLOCK-T3) and warm Camera Hexapod (BLOCK-T4) were completed successfully. Both components were left at enabled state and were monitored (see OBS-592).\n",
-       "\n",
-       "Camera Hexapod Strut12 showed a notorious increase on temperature (about 2C in about 20 min plus diverting in motor current from the rest of the struts). It was turned off and left in standby state. \n",
-       "\n",
-       "For M2 Hexapod, we kept monitoring its currents on enabled state (as temperature sensors are not available, OBS-593). At about 23:30 CLT, motor current on leg 4 approached the 4 amp. We proceed to shutdown the hexapod but it was not possible (either locally nor remote) to access the PDU with the known credentials (IHS-8426).\n",
-       "\n",
-       "The remaining Simonyi tests (BLOCK-T145 M1M3 Hardpoints Breakaway Test and BLOCK-T144 M1M3 Bump Test) run successfully. Actuators [218, 238, 409] FAILED the bump test. \n",
-       "\n",
-       "\n",
-       "\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:46:16 |
We started from Simonyi Hexapods Startup (BLOCK-T151) around 18:30 UTC. 
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:46:16 |
During daytime, Petr K. worked on M1M3 to run bump tests, raise the mirror, and test to find the reference/center position for static support centering (see more details and results at #m1m3_worklog channel). \n",
-       "\n",
-       "We started the Simonyi Hexapods Startup (BLOCK-T151) around 18:30 UTC and it was done without any issues. Then warm-up M2 Hexapod (BLOCK-T3) and warm Camera Hexapod (BLOCK-T4) were completed successfully. Both components were left at enabled state and were monitored (see OBS-592).\n",
-       "\n",
-       "Camera Hexapod Strut12 showed a notorious increase on temperature (about 2C in about 20 min plus diverting in motor current from the rest of the struts). It was turned off and left in standby state. \n",
-       "\n",
-       "For M2 Hexapod, we kept monitoring its currents (as temperature sensors are not available, OBS-593).\n",
-       "\n",
-       "The remaining Simonyi tests (BLOCK-T145 M1M3 Hardpoints Breakaway Test and BLOCK-T144 M1M3 Bump Test) run successfully. Actuators [218, 238, 409] FAILED the bump test. \n",
-       "\n",
-       "\n",
-       "\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:46:16 |
During daytime, Petr K. worked on M1M3 to run bump tests, raise the mirror, and test to find the reference/center position for static support centering (see more details and results at #m1m3_worklog channel). \n",
-       "\n",
-       "We started from Simonyi Hexapods Startup (BLOCK-T151) around 18:30 UTC and it was done without any issues. Then warm-up M2 Hexapod (BLOCK-T3) and warm Camera Hexapod (BLOCK-T4) were completed successfully. \n",
-       "\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:46:16 |
During daytime, Petr K. worked on M1M3 to run bump tests, raise the mirror, and test to find the reference/center position for static support centering (see more details and results at #m1m3_worklog channel). \n",
-       "\n",
-       "We started from Simonyi Hexapods Startup (BLOCK-T151) around 18:30 UTC and it was done without any issues. Then warm-up M2 Hexapod (BLOCK-T3) and warm Camera Hexapod (BLOCK-T4) were completed successfully. Both components were left at enabled state and were monitored.\n",
-       "\n",
-       "Camera Hexapod Strut12 showed a notorious increase on temperature (about 2C in about 20 min plus diverting in motor current from the rest of the struts). It was turned off and left in Standby state. \n",
-       "\n",
-       "M2 Hexapod, we kept monitoring its currents (as temperature sensors are not available).\n",
-       "\n",
-       "BLOCK-T145 M1M3 Hardpoints Breakaway Test run successfully.\n",
-       "\n",
-       "BLOCK-T144 M1M3 Bump Test, Actuators [218, 238, 409] FAILED the bump test. \n",
-       "\n",
-       "\n",
-       "\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:46:16 |
During daytime, Petr K. worked on M1M3 to run bump tests, raise the mirror, and test to find the reference/center position for static support centering (see more details and results at #m1m3_worklog channel). \n",
-       "\n",
-       "We started from Simonyi Hexapods Startup (BLOCK-T151) around 18:30 UTC and it was done without any issues. Then warm-up M2 Hexapod (BLOCK-T3) and warm Camera Hexapod (BLOCK-T4) were completed successfully. Both components were left at enabled state and were monitored.\n",
-       "\n",
-       "Camera Hexapod Strut12 showed a notorious increase on temperature (about 2C in about 20 min plus diverting in motor current from the rest of the struts). It was turned off and left in Standby state. \n",
-       "\n",
-       "M2 Hexapod, we kept monitoring its currents (as temperature sensors is not available)\n",
-       "\n",
-       "\n",
-       "\n",
-       "\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:46:16 |
During daytime, Petr K. worked on M1M3 to run bump tests, raise the mirror, and test to find the reference/center position for static support centering (see more details and results at #m1m3_worklog channel). \n",
-       "\n",
-       "We started the Simonyi Hexapods Startup (BLOCK-T151) around 18:30 UTC and it was done without any issues. Then warm-up M2 Hexapod (BLOCK-T3) and warm Camera Hexapod (BLOCK-T4) were completed successfully. Both components were left at enabled state and were monitored (see OBS-592).\n",
-       "\n",
-       "Camera Hexapod Strut12 showed a notorious increase on temperature (about 2C in about 20 min plus diverting in motor current from the rest of the struts). It was turned off and left in standby state. \n",
-       "\n",
-       "For M2 Hexapod, we kept monitoring its currents on enabled state (as temperature sensors are not available, OBS-593). At about 23:30 CLT, motor current on leg 4 approached the 4 amp. We proceed to shutdown the hexapod but it was not possible (either locally nor remote) to access the PDU with the known credentials (IHS-8426).\n",
-       "\n",
-       "The remaining Simonyi tests (BLOCK-T145 M1M3 Hardpoints Breakaway Test and BLOCK-T144 M1M3 Bump Test) run successfully. Actuators [218, 238, 409] FAILED the bump test. \n",
-       "\n",
-       "\n",
-       "\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 12:46:16 |
"
-      ],
-      "text/plain": [
-       ""
-      ]
-     },
-     "metadata": {},
-     "output_type": "display_data"
-    },
-    {
-     "data": {
-      "text/markdown": [
-       "> 12:46:16 | 
During daytime, Petr K. worked on M1M3 to run bump tests, raise the mirror, and test to find the reference/center position for static support centering (see more details and results at #m1m3_worklog channel). \n",
-       "\n",
-       "We started from Simonyi Hexapods Startup (BLOCK-T151) around 18:30 UTC and it was done without any issues. Then warm-up M2 Hexapod (BLOCK-T3) and warm Camera Hexapod (BLOCK-T4) were completed successfully. \n",
-       "\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - ":EOT" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - } - ], + "outputs": [], "source": [ "# Get data from Night Report log. Display nightly Jira BLOCKS.\n", "nr_adapter = sad.NightReportAdapter(server_url=server,\n", @@ -1341,39 +262,14 @@ }, { "cell_type": "code", - "execution_count": 9, + "execution_count": null, "id": "14", "metadata": { "jupyter": { "source_hidden": true } }, - "outputs": [ - { - "data": { - "text/markdown": [ - "### No Exposure Log records found." - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "Used [API Data](https://summit-lsp.lsst.codes/exposurelog)" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - } - ], + "outputs": [], "source": [ "# Get data from Exposure log. Display time log.\n", "exposure_adapter = sad.ExposurelogAdapter(\n", @@ -1401,48 +297,24 @@ }, { "cell_type": "code", - "execution_count": 10, + "execution_count": null, "id": "15", "metadata": {}, - "outputs": [ - { - "data": { - "text/plain": [ - "0" - ] - }, - "execution_count": 10, - "metadata": {}, - "output_type": "execute_result" - } - ], + "outputs": [], "source": [ "len(exposure_recs)" ] }, { "cell_type": "code", - "execution_count": 11, + "execution_count": null, "id": "16", "metadata": { "jupyter": { "source_hidden": true } }, - "outputs": [ - { - "data": { - "text/markdown": [ - "### No Observation Gaps found in exposures." - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - } - ], + "outputs": [], "source": [ "# Display Observation gaps\n", "if usdf == os.environ.get('EXTERNAL_INSTANCE_URL'):\n", @@ -1471,966 +343,14 @@ }, { "cell_type": "code", - "execution_count": 12, + "execution_count": null, "id": "18", "metadata": { "jupyter": { "source_hidden": true } }, - "outputs": [ - { - "data": { - "text/markdown": [ - "### Warning: Some text of Narrative log message may confuse markdown rendering." - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "## NIGHT: 20240916: " - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "### DATE: 2024-09-17: " - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 02:05:26 |
Resumed AuxTel on-sky\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 02:31:51 |
Output of the Bump Test:\r\n",
-       "2024/09/17 02:26:20 TAI\r\n",
-       "\r\n",
-       "\r\n",
-       "RuntimeError: Actuators [238, 330, 409] FAILED the bump test.\r\n",
-       "\r\n",
-       "\r\n",
-       "SAA (Single Actuator Axes) Failures:\r\n",
-       "  None\r\n",
-       "DAA (Dual Actuator Axes) Failures:\r\n",
-       "  - Actuator ID 238: Pri Index 72\r\n",
-       "  - Actuator ID 238: Sec Index 55\r\n",
-       "  - Actuator ID 330: Pri Index 107\r\n",
-       "  - Actuator ID 330: Pri Index 107\r\n",
-       "  - Actuator ID 409: Pri Index 123\r\n",
-       "  - Actuator ID 409: Pri Index 123\r\n",
-       "\r\n",
-       "\r\n",
-       "\r\n",
-       "\r\n",
-       "\r\n",
-       "\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 02:31:57 |
Tiago is working on Scheduler:1 update.\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 02:57:00 |
Tiago is working on Scheduler:1 update.& ScriptQueue:1\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 04:07:45 |
ATPtg went to FAULT, message 'Azimuth out of range error''\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 04:13:21 |
After recovering ATPtg the scheduler will not resume.\r\n",
-       "\r\n",
-       "\r\n",
-       "SALindex: 200103 - TAI: 4:10 \r\n",
-       "\r\n",
-       "\r\n",
-       "Error in run\r\n",
-       "Traceback (most recent call last):\r\n",
-       "  File \"/opt/lsst/software/stack/conda/envs/lsst-scipipe-9.0.0/lib/python3.11/site-packages/lsst/ts/salobj/base_script.py\", line 631, in do_run\r\n",
-       "    await self._run_task\r\n",
-       "  File \"/net/obs-env/auto_base_packages/ts_standardscripts/python/lsst/ts/standardscripts/scheduler/resume.py\", line 89, in run\r\n",
-       "    await self.scheduler_remote.cmd_resume.set_start(timeout=self.timeout_start)\r\n",
-       "  File \"/opt/lsst/software/stack/conda/envs/lsst-scipipe-9.0.0/lib/python3.11/site-packages/lsst/ts/salobj/topics/remote_command.py\", line 416, in set_start\r\n",
-       "    return await self.start(timeout=timeout, wait_done=wait_done)\r\n",
-       "           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/conda/envs/lsst-scipipe-9.0.0/lib/python3.11/site-packages/lsst/ts/salobj/topics/remote_command.py\", line 487, in start\r\n",
-       "    return await cmd_info.next_ackcmd(timeout=timeout)\r\n",
-       "           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/conda/envs/lsst-scipipe-9.0.0/lib/python3.11/site-packages/lsst/ts/salobj/topics/remote_command.py\", line 191, in next_ackcmd\r\n",
-       "    raise base.AckError(msg=\"Command failed\", ackcmd=ackcmd)\r\n",
-       "lsst.ts.salobj.base.AckError: msg='Command failed', ackcmd=(ackcmd private_seqNum=1364691750, ack=<SalRetCode.CMD_FAILED: -302>, error=1, result='Failed: Detailed state must be IDLE, currently in <DetailedState.RUNNING: 2>.')\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 04:15:42 |
We changed  the scheduler:2 state to DISABLED and then ENABLED from the CSC summary.\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 04:21:25 |
We changed  the scheduler:2 state to DISABLED and then ENABLED from the CSC summary and then we resumed the observations.\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 04:26:44 |
Hexapod strut temperatures have increased about 1.5 C (average) since both warm-up finished. Both Hexapods are DISABLED.\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 04:44:09 |
BLOCK-T139 (1.0) Simonyi Hexapod Shutdown completed\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 05:01:18 |
Dome and telescope in the parking position, ATTCS and LATISS in Standby.\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "## NIGHT: 20240917: " - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "### DATE: 2024-09-17: " - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 15:25:39 |
Vent started from 11:30 am CLT with vent.py. \r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 17:36:34 |
Stop vent.py. Starting Daytime calibration. \r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 18:32:39 |
Starting BLOCK-T151- Simonyi Hexapods Startup \r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 18:54:47 |
Starting BLOCK-T3 (Warm-up M2 Hexapod \r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 18:55:05 |
Starting BLOCK-T3 Warm-up M2 Hexapod \r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 19:04:19 |
Starting BLOCK-T4 - Warm-up Camera Hexapod\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 19:07:04 |
Starting BLOCK-T151- Simonyi Hexapods Startup.\r\n",
-       "\r\n",
-       "\r\n",
-       "Camera hexapod temperature before started was 17.2 degrees  C\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 19:08:19 |
Starting BLOCK-T151- Simonyi Hexapods Startup.\r\n",
-       "\r\n",
-       "\r\n",
-       "Camera hexapod temperature was 17.2 degrees  C and begin to increase as soon  as it was startup.\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 19:22:21 |
MT Hexapod went to FAUlt 19:14 TAI. \r\n",
-       "'Force actuator Y 338 measured force (-372.00507 N) outside limits'\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 21:06:39 |
Starting BLOCK-T153 Opening M1M3 Mirror Covers\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 21:11:02 |
MTMount did not transition to Enabled state.  \r\n",
-       "\r\n",
-       "\r\n",
-       "'begin_enable failed; remaining in state <State.DISABLED: 1>: The CSC was not allowed to command the mount: ExpectedError('Not connected to the low-level controller.')'\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 21:15:35 |
TMA was in EUI mode, that's why we couldn't transition states\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 21:20:10 |
MTM2 went to FAULT state \r\n",
-       "\r\n",
-       "\r\n",
-       "'Lost the TCP/IP connection.'\r\n",
-       "\r\n",
-       "\r\n",
-       "Error happens in the connection request.\r\n",
-       "Traceback (most recent call last):\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/m2com/tcp_client.py\", line 127, in connect\r\n",
-       "    await asyncio.wait_for(self.start_task, timeout=timeout)\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/asyncio/tasks.py\", line 489, in wait_for\r\n",
-       "    return fut.result()\r\n",
-       "           ^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/tcpip/client.py\", line 186, in start\r\n",
-       "    reader, writer = await asyncio.open_connection(\r\n",
-       "                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/asyncio/streams.py\", line 48, in open_connection\r\n",
-       "    transport, _ = await loop.create_connection(\r\n",
-       "                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/asyncio/base_events.py\", line 1086, in create_connection\r\n",
-       "    raise exceptions[0]\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/asyncio/base_events.py\", line 1070, in create_connection\r\n",
-       "    sock = await self._connect_sock(\r\n",
-       "           ^^^^^^^^^^^^^^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/asyncio/base_events.py\", line 974, in _connect_sock\r\n",
-       "    await self.sock_connect(sock, address)\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/asyncio/selector_events.py\", line 638, in sock_connect\r\n",
-       "    return await fut\r\n",
-       "           ^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/asyncio/selector_events.py\", line 678, in _sock_connect_cb\r\n",
-       "    raise OSError(err, f'Connect call failed {address}')\r\n",
-       "ConnectionRefusedError: [Errno 111] Connect call failed ('139.229.178.194', 50011)\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 21:22:05 |
MTM2 is now ENABLED\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 21:30:05 |
BLOCK-T152  Closing M1M3 Mirror Covers run succesfully\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 21:30:26 |
Running BLOCK-T145 M1M3 Hardpoints Breakaway\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 21:34:19 |
Camera hexapod strut 12 temperature is increasing rapidly unlike other temperatures \r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 21:37:22 |
Camera hexapod strut 12 temperature is rising rapidly, unlike the other strut temperature, which are increasing more gradually.\r\n",
-       "\r\n",
-       "\r\n",
-       "after warmup it was 17 degrees C  and it reached  19.3 degrees C  \r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 21:39:08 |
Camera hexapod shutdown and set to STANDBY state\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 22:40:57 |
Scheduler:2 went to FAULT state \r\n",
-       "\r\n",
-       "\r\n",
-       "Error on advance target production loop.\r\n",
-       "Traceback (most recent call last):\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 1463, in advance_target_production_loop\r\n",
-       "    await self.generate_target_queue()\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/utils/csc_utils.py\", line 200, in detailed_state_wrapper\r\n",
-       "    await coroutine(self, *args, **kwargs)\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 1641, in generate_target_queue\r\n",
-       "    await self.handle_no_targets_on_queue()\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 1733, in handle_no_targets_on_queue\r\n",
-       "    await self.put_on_queue([stop_tracking_target])\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 872, in put_on_queue\r\n",
-       "    async for sal_index in self._queue_block_scripts(observing_block):\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 928, in _queue_block_scripts\r\n",
-       "    sal_index = await self._queue_one_script(\r\n",
-       "                ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 984, in _queue_one_script\r\n",
-       "    add_task = await self.queue_remote.cmd_add.set_start(\r\n",
-       "               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/salobj/topics/remote_command.py\", line 416, in set_start\r\n",
-       "    return await self.start(timeout=timeout, wait_done=wait_done)\r\n",
-       "           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/salobj/topics/remote_command.py\", line 487, in start\r\n",
-       "    return await cmd_info.next_ackcmd(timeout=timeout)\r\n",
-       "           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/salobj/topics/remote_command.py\", line 191, in next_ackcmd\r\n",
-       "    raise base.AckError(msg=\"Command failed\", ackcmd=ackcmd)\r\n",
-       "lsst.ts.salobj.base.AckError: msg='Command failed', ackcmd=(ackcmd private_seqNum=78556861, ack=<SalRetCode.CMD_FAILED: -302>, error=1, result='Failed: _Internal has the wrong format, should be BLOCK-N or BLOCK-TN.')\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 22:50:18 |
Scheduler:2 went to FAULT state after resume.py 3 times, after enable it from ATQueue.\r\n",
-       "\r\n",
-       "\r\n",
-       "Error on advance target production loop.\r\n",
-       "Traceback (most recent call last):\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 1463, in advance_target_production_loop\r\n",
-       "    await self.generate_target_queue()\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/utils/csc_utils.py\", line 200, in detailed_state_wrapper\r\n",
-       "    await coroutine(self, *args, **kwargs)\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 1641, in generate_target_queue\r\n",
-       "    await self.handle_no_targets_on_queue()\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 1733, in handle_no_targets_on_queue\r\n",
-       "    await self.put_on_queue([stop_tracking_target])\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 872, in put_on_queue\r\n",
-       "    async for sal_index in self._queue_block_scripts(observing_block):\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 928, in _queue_block_scripts\r\n",
-       "    sal_index = await self._queue_one_script(\r\n",
-       "                ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 984, in _queue_one_script\r\n",
-       "    add_task = await self.queue_remote.cmd_add.set_start(\r\n",
-       "               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/salobj/topics/remote_command.py\", line 416, in set_start\r\n",
-       "    return await self.start(timeout=timeout, wait_done=wait_done)\r\n",
-       "           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/salobj/topics/remote_command.py\", line 487, in start\r\n",
-       "    return await cmd_info.next_ackcmd(timeout=timeout)\r\n",
-       "           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/salobj/topics/remote_command.py\", line 191, in next_ackcmd\r\n",
-       "    raise base.AckError(msg=\"Command failed\", ackcmd=ackcmd)\r\n",
-       "lsst.ts.salobj.base.AckError: msg='Command failed', ackcmd=(ackcmd private_seqNum=78556861, ack=<SalRetCode.CMD_FAILED: -302>, error=1, result='Failed: _Internal has the wrong format, should be BLOCK-N or BLOCK-TN.')\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 22:52:36 |
Scheduler:2 went to FAULT state after resume.py 3 times, after enable it from ATQueue. This is currently a bug in the update  that needs to be fix.\r\n",
-       "\r\n",
-       "\r\n",
-       "Error on advance target production loop.\r\n",
-       "Traceback (most recent call last):\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 1463, in advance_target_production_loop\r\n",
-       "    await self.generate_target_queue()\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/utils/csc_utils.py\", line 200, in detailed_state_wrapper\r\n",
-       "    await coroutine(self, *args, **kwargs)\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 1641, in generate_target_queue\r\n",
-       "    await self.handle_no_targets_on_queue()\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 1733, in handle_no_targets_on_queue\r\n",
-       "    await self.put_on_queue([stop_tracking_target])\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 872, in put_on_queue\r\n",
-       "    async for sal_index in self._queue_block_scripts(observing_block):\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 928, in _queue_block_scripts\r\n",
-       "    sal_index = await self._queue_one_script(\r\n",
-       "                ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 984, in _queue_one_script\r\n",
-       "    add_task = await self.queue_remote.cmd_add.set_start(\r\n",
-       "               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/salobj/topics/remote_command.py\", line 416, in set_start\r\n",
-       "    return await self.start(timeout=timeout, wait_done=wait_done)\r\n",
-       "           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/salobj/topics/remote_command.py\", line 487, in start\r\n",
-       "    return await cmd_info.next_ackcmd(timeout=timeout)\r\n",
-       "           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/salobj/topics/remote_command.py\", line 191, in next_ackcmd\r\n",
-       "    raise base.AckError(msg=\"Command failed\", ackcmd=ackcmd)\r\n",
-       "lsst.ts.salobj.base.AckError: msg='Command failed', ackcmd=(ackcmd private_seqNum=78556861, ack=<SalRetCode.CMD_FAILED: -302>, error=1, result='Failed: _Internal has the wrong format, should be BLOCK-N or BLOCK-TN.')\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 22:54:57 |
Scheduler:2 went to FAULT state in the  resume.py 3 times, after enable it from ATQueue. This is currently a bug in the shceduler:2 update that needs to be fix. The solution is to wait until there are targets available.\r\n",
-       "\r\n",
-       "\r\n",
-       "First occurrence  TAI:  22:39 (the script did not fault)\r\n",
-       "\r\n",
-       "\r\n",
-       "Error on advance target production loop.\r\n",
-       "Traceback (most recent call last):\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 1463, in advance_target_production_loop\r\n",
-       "    await self.generate_target_queue()\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/utils/csc_utils.py\", line 200, in detailed_state_wrapper\r\n",
-       "    await coroutine(self, *args, **kwargs)\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 1641, in generate_target_queue\r\n",
-       "    await self.handle_no_targets_on_queue()\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 1733, in handle_no_targets_on_queue\r\n",
-       "    await self.put_on_queue([stop_tracking_target])\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 872, in put_on_queue\r\n",
-       "    async for sal_index in self._queue_block_scripts(observing_block):\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 928, in _queue_block_scripts\r\n",
-       "    sal_index = await self._queue_one_script(\r\n",
-       "                ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 984, in _queue_one_script\r\n",
-       "    add_task = await self.queue_remote.cmd_add.set_start(\r\n",
-       "               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/salobj/topics/remote_command.py\", line 416, in set_start\r\n",
-       "    return await self.start(timeout=timeout, wait_done=wait_done)\r\n",
-       "           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/salobj/topics/remote_command.py\", line 487, in start\r\n",
-       "    return await cmd_info.next_ackcmd(timeout=timeout)\r\n",
-       "           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/salobj/topics/remote_command.py\", line 191, in next_ackcmd\r\n",
-       "    raise base.AckError(msg=\"Command failed\", ackcmd=ackcmd)\r\n",
-       "lsst.ts.salobj.base.AckError: msg='Command failed', ackcmd=(ackcmd private_seqNum=78556861, ack=<SalRetCode.CMD_FAILED: -302>, error=1, result='Failed: _Internal has the wrong format, should be BLOCK-N or BLOCK-TN.')\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 23:22:41 |
We are running BLOCK-T144 M1M3 Bump Test\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 00:02:15 |
AuxTel is closed due to bad weather conditions\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 00:28:44 |
Actuators [218, 238, 409] FAILED the bump test.\r\n",
-       "\r\n",
-       "\r\n",
-       "SAA (Single Actuator Axes) Failures:\r\n",
-       "  None\r\n",
-       "DAA (Dual Actuator Axes) Failures:\r\n",
-       "  - Actuator ID 218: Pri Index 53\r\n",
-       "  - Actuator ID 218: Pri Index 53\r\n",
-       "  - Actuator ID 238: Pri Index 72\r\n",
-       "  - Actuator ID 238: Sec Index 55\r\n",
-       "  - Actuator ID 409: Pri Index 123\r\n",
-       "  - Actuator ID 409: Pri Index 123\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 00:32:46 |
Scheduler:2 went to FAULT state during the  resume.py 3 times, after enable it from ATQueue. This is currently a bug in the scheduler:2 update.The solution is to wait until there are targets available.\r\n",
-       "\r\n",
-       "\r\n",
-       "First occurrence  TAI:  22:39 (the script did not fault)\r\n",
-       "\r\n",
-       "\r\n",
-       "Error on advance target production loop.\r\n",
-       "Traceback (most recent call last):\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 1463, in advance_target_production_loop\r\n",
-       "    await self.generate_target_queue()\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/utils/csc_utils.py\", line 200, in detailed_state_wrapper\r\n",
-       "    await coroutine(self, *args, **kwargs)\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 1641, in generate_target_queue\r\n",
-       "    await self.handle_no_targets_on_queue()\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 1733, in handle_no_targets_on_queue\r\n",
-       "    await self.put_on_queue([stop_tracking_target])\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 872, in put_on_queue\r\n",
-       "    async for sal_index in self._queue_block_scripts(observing_block):\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 928, in _queue_block_scripts\r\n",
-       "    sal_index = await self._queue_one_script(\r\n",
-       "                ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/scheduler/scheduler_csc.py\", line 984, in _queue_one_script\r\n",
-       "    add_task = await self.queue_remote.cmd_add.set_start(\r\n",
-       "               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/salobj/topics/remote_command.py\", line 416, in set_start\r\n",
-       "    return await self.start(timeout=timeout, wait_done=wait_done)\r\n",
-       "           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/salobj/topics/remote_command.py\", line 487, in start\r\n",
-       "    return await cmd_info.next_ackcmd(timeout=timeout)\r\n",
-       "           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n",
-       "  File \"/opt/lsst/software/stack/miniconda/lib/python3.11/site-packages/lsst/ts/salobj/topics/remote_command.py\", line 191, in next_ackcmd\r\n",
-       "    raise base.AckError(msg=\"Command failed\", ackcmd=ackcmd)\r\n",
-       "lsst.ts.salobj.base.AckError: msg='Command failed', ackcmd=(ackcmd private_seqNum=78556861, ack=<SalRetCode.CMD_FAILED: -302>, error=1, result='Failed: _Internal has the wrong format, should be BLOCK-N or BLOCK-TN.')\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 00:37:13 |
M1M3 went to FAULT state 'Force actuator Y 338 measured force (-372.40256 N) outside limits'\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 00:52:10 |
Preparing to go on sky\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 02:58:53 |
M2 hexapod set to standby state\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 03:18:14 |
We could not login to  https://pdu1-tea-as02.cp.lsst.org to turn off the drives and PXI. \r\n",
-       "Even though we were able to do so earlier.\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 03:31:55 |
M2 Hexapod set to disabled state\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 03:33:57 |
M2 Hexapod set to disabled state. \r\n",
-       "Currents now  is around 0 \r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 03:38:43 |
M2 Hexapod set to disabled state. \r\n",
-       "Currents now are around 0 \r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 04:41:06 |
Stopping observations and will begin shutdown\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 04:53:33 |
AuxTel Dome and Telescope at parking position.\r\n",
-       "ATTCS and LATISS components in Standby.\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "> 04:54:04 |
Final status: \r\n",
-       "MTMount, M1M3 disabled.\r\n",
-       "MTPtg & MTM2 enabled.\r\n",
-       "Camera Hexapod in Standby state with drives and PXI turned off.\r\n",
-       "M2 Hexapod in Disabled state with drives and PXI turned on.\r\n",
-       "
" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - ":EOT" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - } - ], + "outputs": [], "source": [ "# Get data from Narrative log. Display time log.\n", "narrative_adapter = sad.NarrativelogAdapter(\n", @@ -2468,72 +388,10 @@ }, { "cell_type": "code", - "execution_count": 13, + "execution_count": null, "id": "19", "metadata": {}, - "outputs": [ - { - "data": { - "text/markdown": [ - "# Dashboard" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/markdown": [ - "(This is not done when running under Times Square.)" - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Try connect to each endpoint of https://summit-lsp.lsst.codes/exposurelog \n", - "Try connect to each endpoint of https://summit-lsp.lsst.codes/narrativelog \n", - "Try connect to each endpoint of https://summit-lsp.lsst.codes/nightreport \n", - "Try connect to each endpoint of https://usdf-rsp-dev.slac.stanford.edu/exposurelog \n", - "Try connect to each endpoint of https://usdf-rsp-dev.slac.stanford.edu/narrativelog \n", - "Try connect to each endpoint of https://usdf-rsp-dev.slac.stanford.edu/nightreport \n", - "Try connect to each endpoint of https://tucson-teststand.lsst.codes/exposurelog \n", - "Try connect to each endpoint of https://tucson-teststand.lsst.codes/narrativelog \n", - "Try connect to each endpoint of https://tucson-teststand.lsst.codes/nightreport \n", - "\n", - "Connected to 9 out of 15 endpoints.(60%)\n", - "Successful connects (9): \n", - "\thttps://summit-lsp.lsst.codes/exposurelog/instruments\n", - "\thttps://summit-lsp.lsst.codes/exposurelog/exposures?instrument=na\n", - "\thttps://summit-lsp.lsst.codes/exposurelog/messages\n", - "\thttps://summit-lsp.lsst.codes/narrativelog/messages\n", - "\thttps://summit-lsp.lsst.codes/nightreport/reports\n", - "\thttps://usdf-rsp-dev.slac.stanford.edu/exposurelog/messages\n", - "\thttps://usdf-rsp-dev.slac.stanford.edu/narrativelog/messages\n", - "\thttps://tucson-teststand.lsst.codes/exposurelog/instruments\n", - "\thttps://tucson-teststand.lsst.codes/exposurelog/exposures?instrument=na\n", - "Failed connects (6): \n", - "\t500: https://usdf-rsp-dev.slac.stanford.edu/exposurelog/instruments\n", - "\t500: https://usdf-rsp-dev.slac.stanford.edu/exposurelog/exposures?instrument=na\n", - "\t404: https://usdf-rsp-dev.slac.stanford.edu/nightreport/reports\n", - "\t500: https://tucson-teststand.lsst.codes/exposurelog/messages\n", - "\t500: https://tucson-teststand.lsst.codes/narrativelog/messages\n", - "\t500: https://tucson-teststand.lsst.codes/nightreport/reports\n", - "score=60%\n", - "Servers that are fully functional for Logging and Reporting:\n", - "\t https://summit-lsp.lsst.codes\n" - ] - } - ], + "outputs": [], "source": [ "# Conditionally display our current ability to connect to all needed endpoints.\n", "if not os.environ.get('EXTERNAL_INSTANCE_URL'):\n", @@ -2544,98 +402,20 @@ }, { "cell_type": "code", - "execution_count": 14, + "execution_count": null, "id": "20", "metadata": {}, - "outputs": [ - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Finished 2024-09-18 14:53:49.364756\n" - ] - } - ], + "outputs": [], "source": [ "print(f'Finished {str(datetime.now())}')" ] }, { "cell_type": "code", - "execution_count": 16, + "execution_count": null, "id": "21", "metadata": {}, - "outputs": [ - { - "name": "stdout", - "output_type": "stream", - "text": [ - "/Users/vbecker/Develop/lsst/ts_repos/ts_logging_and_reporting/notebooks_tsqr\n", - "['NightLog.ipynb', 'logrep_proto_1.ipynb', 'requirements.txt', 'NightLog.yaml', 'efd.ipynb', 'NightLog2.ipynb', 'scaffolding.org', 'README.md', 'dashboard.ipynb', 'exposurelog.HIDE_yaml', 'consdb', '.ipynb_checkpoints', 'TEMPLATE_logrep.HIDE_yaml', 'logrep_all_env.ipynb', 'narrativelog.HIDE_yaml', 'efd.HIDE_yaml', 'narrativelog.ipynb', 'exposurelog.ipynb', 'TEMPLATE_logrep.ipynb']\n", - "None\n", - "Installing \"lsst.sitcom.summit.utils\" from github using \"main\" branch....\n", - " Running command git clone --filter=blob:none --quiet https://github.com/lsst-sitcom/summit_utils.git /private/var/folders/s5/114qkzh55t36hvzbb6blzdq80000gp/T/pip-req-build-tpomyp4v\n", - " \u001b[1;31merror\u001b[0m: \u001b[1msubprocess-exited-with-error\u001b[0m\n", - " \n", - " \u001b[31m×\u001b[0m \u001b[32mGetting requirements to build wheel\u001b[0m did not run successfully.\n", - " \u001b[31m│\u001b[0m exit code: \u001b[1;36m1\u001b[0m\n", - " \u001b[31m╰─>\u001b[0m \u001b[31m[14 lines of output]\u001b[0m\n", - " \u001b[31m \u001b[0m error: Multiple top-level packages discovered in a flat-layout: ['ups', 'config', 'pipelines'].\n", - " \u001b[31m \u001b[0m \n", - " \u001b[31m \u001b[0m To avoid accidental inclusion of unwanted files or directories,\n", - " \u001b[31m \u001b[0m setuptools will not proceed with this build.\n", - " \u001b[31m \u001b[0m \n", - " \u001b[31m \u001b[0m If you are trying to create a single distribution with multiple packages\n", - " \u001b[31m \u001b[0m on purpose, you should not rely on automatic discovery.\n", - " \u001b[31m \u001b[0m Instead, consider the following options:\n", - " \u001b[31m \u001b[0m \n", - " \u001b[31m \u001b[0m 1. set up custom discovery (`find` directive with `include` or `exclude`)\n", - " \u001b[31m \u001b[0m 2. use a `src-layout`\n", - " \u001b[31m \u001b[0m 3. explicitly set `py_modules` or `packages` with a list of names\n", - " \u001b[31m \u001b[0m \n", - " \u001b[31m \u001b[0m To find more information, look for \"package discovery\" on setuptools docs.\n", - " \u001b[31m \u001b[0m \u001b[31m[end of output]\u001b[0m\n", - " \n", - " \u001b[1;35mnote\u001b[0m: This error originates from a subprocess, and is likely not a problem with pip.\n", - "\u001b[1;31merror\u001b[0m: \u001b[1msubprocess-exited-with-error\u001b[0m\n", - "\n", - "\u001b[31m×\u001b[0m \u001b[32mGetting requirements to build wheel\u001b[0m did not run successfully.\n", - "\u001b[31m│\u001b[0m exit code: \u001b[1;36m1\u001b[0m\n", - "\u001b[31m╰─>\u001b[0m See above for output.\n", - "\n", - "\u001b[1;35mnote\u001b[0m: This error originates from a subprocess, and is likely not a problem with pip.\n", - "Note: you may need to restart the kernel to use updated packages.\n" - ] - }, - { - "ename": "ModuleNotFoundError", - "evalue": "No module named 'lsst.summit'", - "output_type": "error", - "traceback": [ - "\u001b[0;31m---------------------------------------------------------------------------\u001b[0m", - "\u001b[0;31mModuleNotFoundError\u001b[0m Traceback (most recent call last)", - "File \u001b[0;32m/var/folders/s5/114qkzh55t36hvzbb6blzdq80000gp/T/ipykernel_58071/917285549.py:8\u001b[0m\n\u001b[1;32m 5\u001b[0m \u001b[38;5;28mprint\u001b[39m(\u001b[38;5;124m'\u001b[39m\u001b[38;5;124mInstalling \u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mlsst.sitcom.summit.utils\u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124m from github using \u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mmain\u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124m branch....\u001b[39m\u001b[38;5;124m'\u001b[39m)\n\u001b[1;32m 6\u001b[0m get_ipython()\u001b[38;5;241m.\u001b[39mrun_line_magic(\u001b[38;5;124m'\u001b[39m\u001b[38;5;124mpip\u001b[39m\u001b[38;5;124m'\u001b[39m, \u001b[38;5;124m'\u001b[39m\u001b[38;5;124minstall --upgrade git+https://github.com/lsst-sitcom/summit_utils.git@main>/dev/null\u001b[39m\u001b[38;5;124m'\u001b[39m)\n\u001b[0;32m----> 8\u001b[0m \u001b[38;5;28;01mfrom\u001b[39;00m \u001b[38;5;21;01mlsst\u001b[39;00m\u001b[38;5;21;01m.\u001b[39;00m\u001b[38;5;21;01msummit\u001b[39;00m\u001b[38;5;21;01m.\u001b[39;00m\u001b[38;5;21;01mutils\u001b[39;00m \u001b[38;5;28;01mimport\u001b[39;00m ConsDbClient\n", - "\u001b[0;31mModuleNotFoundError\u001b[0m: No module named 'lsst.summit'" - ] - }, - { - "ename": "ModuleNotFoundError", - "evalue": "No module named 'lsst.summit'", - "output_type": "error", - "traceback": [ - "\u001b[0;31m---------------------------------------------------------------------------\u001b[0m", - "\u001b[0;31mModuleNotFoundError\u001b[0m Traceback (most recent call last)", - "Cell \u001b[0;32mIn[16], line 3\u001b[0m\n\u001b[1;32m 1\u001b[0m \u001b[38;5;28mprint\u001b[39m(os\u001b[38;5;241m.\u001b[39mgetcwd())\n\u001b[1;32m 2\u001b[0m \u001b[38;5;28mprint\u001b[39m(os\u001b[38;5;241m.\u001b[39mlistdir())\n\u001b[0;32m----> 3\u001b[0m get_ipython()\u001b[38;5;241m.\u001b[39mrun_line_magic(\u001b[38;5;124m'\u001b[39m\u001b[38;5;124mrun\u001b[39m\u001b[38;5;124m'\u001b[39m, \u001b[38;5;124m'\u001b[39m\u001b[38;5;124mconsdb/assorted_plots.ipynb\u001b[39m\u001b[38;5;124m'\u001b[39m)\n", - "File \u001b[0;32m/opt/anaconda3/envs/logrep/lib/python3.12/site-packages/IPython/core/interactiveshell.py:2480\u001b[0m, in \u001b[0;36mInteractiveShell.run_line_magic\u001b[0;34m(self, magic_name, line, _stack_depth)\u001b[0m\n\u001b[1;32m 2478\u001b[0m kwargs[\u001b[38;5;124m'\u001b[39m\u001b[38;5;124mlocal_ns\u001b[39m\u001b[38;5;124m'\u001b[39m] \u001b[38;5;241m=\u001b[39m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39mget_local_scope(stack_depth)\n\u001b[1;32m 2479\u001b[0m \u001b[38;5;28;01mwith\u001b[39;00m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39mbuiltin_trap:\n\u001b[0;32m-> 2480\u001b[0m result \u001b[38;5;241m=\u001b[39m fn(\u001b[38;5;241m*\u001b[39margs, \u001b[38;5;241m*\u001b[39m\u001b[38;5;241m*\u001b[39mkwargs)\n\u001b[1;32m 2482\u001b[0m \u001b[38;5;66;03m# The code below prevents the output from being displayed\u001b[39;00m\n\u001b[1;32m 2483\u001b[0m \u001b[38;5;66;03m# when using magics with decorator @output_can_be_silenced\u001b[39;00m\n\u001b[1;32m 2484\u001b[0m \u001b[38;5;66;03m# when the last Python token in the expression is a ';'.\u001b[39;00m\n\u001b[1;32m 2485\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m \u001b[38;5;28mgetattr\u001b[39m(fn, magic\u001b[38;5;241m.\u001b[39mMAGIC_OUTPUT_CAN_BE_SILENCED, \u001b[38;5;28;01mFalse\u001b[39;00m):\n", - "File \u001b[0;32m/opt/anaconda3/envs/logrep/lib/python3.12/site-packages/IPython/core/magics/execution.py:737\u001b[0m, in \u001b[0;36mExecutionMagics.run\u001b[0;34m(self, parameter_s, runner, file_finder)\u001b[0m\n\u001b[1;32m 735\u001b[0m \u001b[38;5;28;01mwith\u001b[39;00m preserve_keys(\u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39mshell\u001b[38;5;241m.\u001b[39muser_ns, \u001b[38;5;124m'\u001b[39m\u001b[38;5;124m__file__\u001b[39m\u001b[38;5;124m'\u001b[39m):\n\u001b[1;32m 736\u001b[0m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39mshell\u001b[38;5;241m.\u001b[39muser_ns[\u001b[38;5;124m'\u001b[39m\u001b[38;5;124m__file__\u001b[39m\u001b[38;5;124m'\u001b[39m] \u001b[38;5;241m=\u001b[39m filename\n\u001b[0;32m--> 737\u001b[0m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39mshell\u001b[38;5;241m.\u001b[39msafe_execfile_ipy(filename, raise_exceptions\u001b[38;5;241m=\u001b[39m\u001b[38;5;28;01mTrue\u001b[39;00m)\n\u001b[1;32m 738\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m\n\u001b[1;32m 740\u001b[0m \u001b[38;5;66;03m# Control the response to exit() calls made by the script being run\u001b[39;00m\n", - "File \u001b[0;32m/opt/anaconda3/envs/logrep/lib/python3.12/site-packages/IPython/core/interactiveshell.py:3005\u001b[0m, in \u001b[0;36mInteractiveShell.safe_execfile_ipy\u001b[0;34m(self, fname, shell_futures, raise_exceptions)\u001b[0m\n\u001b[1;32m 3003\u001b[0m result \u001b[38;5;241m=\u001b[39m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39mrun_cell(cell, silent\u001b[38;5;241m=\u001b[39m\u001b[38;5;28;01mTrue\u001b[39;00m, shell_futures\u001b[38;5;241m=\u001b[39mshell_futures)\n\u001b[1;32m 3004\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m raise_exceptions:\n\u001b[0;32m-> 3005\u001b[0m result\u001b[38;5;241m.\u001b[39mraise_error()\n\u001b[1;32m 3006\u001b[0m \u001b[38;5;28;01melif\u001b[39;00m \u001b[38;5;129;01mnot\u001b[39;00m result\u001b[38;5;241m.\u001b[39msuccess:\n\u001b[1;32m 3007\u001b[0m \u001b[38;5;28;01mbreak\u001b[39;00m\n", - "File \u001b[0;32m/opt/anaconda3/envs/logrep/lib/python3.12/site-packages/IPython/core/interactiveshell.py:308\u001b[0m, in \u001b[0;36mExecutionResult.raise_error\u001b[0;34m(self)\u001b[0m\n\u001b[1;32m 306\u001b[0m \u001b[38;5;28;01mraise\u001b[39;00m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39merror_before_exec\n\u001b[1;32m 307\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39merror_in_exec \u001b[38;5;129;01mis\u001b[39;00m \u001b[38;5;129;01mnot\u001b[39;00m \u001b[38;5;28;01mNone\u001b[39;00m:\n\u001b[0;32m--> 308\u001b[0m \u001b[38;5;28;01mraise\u001b[39;00m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39merror_in_exec\n", - " \u001b[0;31m[... skipping hidden 1 frame]\u001b[0m\n", - "File \u001b[0;32m/var/folders/s5/114qkzh55t36hvzbb6blzdq80000gp/T/ipykernel_58071/917285549.py:8\u001b[0m\n\u001b[1;32m 5\u001b[0m \u001b[38;5;28mprint\u001b[39m(\u001b[38;5;124m'\u001b[39m\u001b[38;5;124mInstalling \u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mlsst.sitcom.summit.utils\u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124m from github using \u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mmain\u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124m branch....\u001b[39m\u001b[38;5;124m'\u001b[39m)\n\u001b[1;32m 6\u001b[0m get_ipython()\u001b[38;5;241m.\u001b[39mrun_line_magic(\u001b[38;5;124m'\u001b[39m\u001b[38;5;124mpip\u001b[39m\u001b[38;5;124m'\u001b[39m, \u001b[38;5;124m'\u001b[39m\u001b[38;5;124minstall --upgrade git+https://github.com/lsst-sitcom/summit_utils.git@main>/dev/null\u001b[39m\u001b[38;5;124m'\u001b[39m)\n\u001b[0;32m----> 8\u001b[0m \u001b[38;5;28;01mfrom\u001b[39;00m \u001b[38;5;21;01mlsst\u001b[39;00m\u001b[38;5;21;01m.\u001b[39;00m\u001b[38;5;21;01msummit\u001b[39;00m\u001b[38;5;21;01m.\u001b[39;00m\u001b[38;5;21;01mutils\u001b[39;00m \u001b[38;5;28;01mimport\u001b[39;00m ConsDbClient\n", - "\u001b[0;31mModuleNotFoundError\u001b[0m: No module named 'lsst.summit'" - ] - } - ], + "outputs": [], "source": [ "print(os.getcwd())\n", "print(os.listdir())\n", @@ -2645,7 +425,7 @@ { "cell_type": "code", "execution_count": null, - "id": "21", + "id": "22", "metadata": {}, "outputs": [], "source": []