Expand blog automation to include information from Google Form (author name etc) as well as copying the image file to our repo#644
Conversation
…pdate info from csv, including downloading the blog image
|
I was thinking about the trigger we need to consider. Actually we have to publish the blog after it's been reviewed and approved. In this case we can introduce two columns: |
| @@ -1,23 +1,280 @@ | |||
| affine @ file:///home/conda/feedstock_root/build_artifacts/affine_1733762038348/work | |||
There was a problem hiding this comment.
Most entries in requirements.txtuse @ file:///... URLs pointing to local build artifact paths on your machine. I think, these paths don't exist when someone else tries to install from them.
There should be a requirements.txt file with just package names and versions.
There was a problem hiding this comment.
And there are soo many dependencies, do we need them all? :)
| ## Run Automation | ||
| 1. Activate virtual environment: `source venv/bin/activate` | ||
| 2. Run the script: `python doc_to_html_conversion.py <DOC_ID>` | ||
| 2. Run the script: `python blog_exporter [--row_index <ROW_INDEX>]`, where the row_index refers to the row of the CSV. This defaults to -1, or the last row in the CSV. |
There was a problem hiding this comment.
Do we need to provide a document id while executing blog_exporter.py?
| ## Run Automation | ||
| 1. Activate virtual environment: `source venv/bin/activate` | ||
| 2. Run the script: `python doc_to_html_conversion.py <DOC_ID>` | ||
| 2. Run the script: `python blog_exporter [--row_index <ROW_INDEX>]`, where the row_index refers to the row of the CSV. This defaults to -1, or the last row in the CSV. |
There was a problem hiding this comment.
| 2. Run the script: `python blog_exporter [--row_index <ROW_INDEX>]`, where the row_index refers to the row of the CSV. This defaults to -1, or the last row in the CSV. | |
| 2. Run the script: `python blog_exporter.py [--row_index <ROW_INDEX>]`, where the row_index refers to the row of the CSV. This defaults to -1, or the last row in the CSV. |
| @@ -0,0 +1,92 @@ | |||
| name: Import Meetup Events | |||
There was a problem hiding this comment.
| name: Import Meetup Events | |
| name: Publish Blogs |
| python -m pip install --upgrade pip | ||
| pip install -r tools/requirements.txt | ||
|
|
||
| - name: Export new blogs |
There was a problem hiding this comment.
For this step I don't see service_account_key.json secret file to be passed into pipeline, will it still work?
|
|
||
|
|
||
|
|
||
|
|
||
|
|
| @@ -0,0 +1,31 @@ | |||
| ,url,doc_id,author_name,author_role,description,source,image_link | |||
There was a problem hiding this comment.
I am not sure we want to expose snapshot publicly...
There was a problem hiding this comment.
Another approach would be - add two columns in excel (isReviewed and isProcessed):
- Read the spreadsheet
- Filter rows where "isReviewed" is yes and "Processed" column is empty
- Export those blogs
- Mark them as "Processed" by writing timestamp back to the sheet
Description
Change Type
Related Issue
#643
Screenshots
Pull request checklist
Please check if your PR fulfills the following requirements: