Template workaround for "When an Excel row is created", "When an Excel row is modified", and/or "When an Excel row is deleted" triggers.
Limitations
• Initially only set up for Excel tables of less than 100,000 records and less than 100MB contents. Version b allows for more than 100,000 records, but still has a 100MB contents limit.
• Each record must be unique. So each record must have a column or combination of columns that is unique & not empty for every single row. In other words, it must include a primary key.
• Every time a new row is added, it must include the primary key column(s) value(s).
• For large tables, the initial set-up may have a 10+ minute delay between the last edits & running whatever actions one adds for the rest of the flow. Version b does include a set-up that is almost twice as fast, but even that will still see a several minute delay after the last edit when the selected Excel file is 10s of thousands of rows.
Initial table
Table edits
Flow run for the 50,000 record table
Example email message with HTML tables for each filtered set of records
HTML table styling: https://ryanmaclean365.com/2020/01/29/power-automate-html-table-styling/
Go to the bottom of the post, download the zip file, go to the page for your flows, and select the legacy import option
Upload the import, change the connections, & select the import button
Select the Open Flow link & delete the initial placeholder value compose action
Switch to your desktop, create a text file on your desktop, input something that follows a JSON structure, go to save the file, add a .json file extension while changing the Save as type to All files & preferably give the file a name that refers to the target Excel workbook name & table name
Go to the same SharePoint library as your Excel workbook with the Excel table you are creating the flow for & upload the new JSON file. Make sure it is set up in a place where no one else will delete or alter it.
In the list for the SharePoint library select the Add column option, select the Show or hide columns option, then select to show the ID column & Apply the changes. Once the ID column is showing, go to the row with the Excel workbook you are creating this for & copy the file row ID.
Switch back to the flow. Open the 1st trigger action "When an item or file is modified" & choose the site & document library for your Excel file.
Then go to the 3 dots on the 1st trigger action "When an item or file is modified", go to the Settings, select the +Add button under Trigger Conditions, input the ID copied for the file row into the expression & paste the expression into the Trigger Conditions & select the Done button
@equals(triggerOutputs()?['body/ID'], InsertIdNumberHere)
Open the 1st Scope of the flow "Only let the flow run if all users stopped editing for a set time", Open the "Get file properties" action, and select the Site address & SharePoint Library Name for your Excel workbook
Then open the "Settings" action & input the name(s) of the column(s) that form a unique key for every row. So the names for the column or combination of columns with values that are unique for every row in the table
Then go to the Scope below that "Read the Excel table and only let the flow run if no edits during read" and in the Excel List rows & Get file properties 2 actions select the Excel Location, Document Library, File, Table, Site Address, & Library Name
If you are using version b, then you will need to input the Excel file & table references in two places:
Then go to the next Scope "Load previous table version copy before updating to the most recent table copy" and in the Get file & Update file actions select the Site Address, File Path, Site Address, & Site Address for the JSON file you loaded into the SharePoint library. This will hold a JSON array copy of the Excel table for the flow runs.
Save & run the flow once so any existing data in your Excel table is recorded in the previous table version JSON file. This will prevent any later actions you add from running for every existing row the 1st time the flow is triggered. By running it once without added actions, it will correctly run your added actions for only the edited rows the 1st time it is triggered later.
That is all the set-up for the template that will run whenever anything in the Excel workbook is modified and that will output different sets (arrays) of the Created records, Modified records & Deleted records. From here you can use the example Apply to each loops & example compose action inputs to set up what actions you want the flow to run for newly created records, for modified records, and/or for deleted records. In the example Apply to each loops the From field is filled with the array Body of the appropriate filter array action for the Created, Modified, or Deleted records and within the loop you can reference any of the record column values with the provided expression by inputting your chosen Apply to each loop name & inputting your chosen column name.
items('Insert_Loop_Name_Here')?['InsertColumnNameHere']
Also if you do not plan to ever use one of the Created, Modified, or Deleted outputs, then you can delete both the filter array action for those records & the associated Apply to each loop
And if you only want the flow actions to run when changes are made to specific columns, you can set the Excel List rows table read to only Select the primary key column & those specific columns to check for changes as described here.
Version B:
Version B replaces the single 100,000 pagination Excel List rows connector with a bit more complicated set-up that determines a rough approximation of the total rows in the table, then loads several 5000 row batches from the table with concurrency/in parallel and then combines all those 5000 row batches into a single JSON array output similar to the output of the standard 100,000 pagination Excel list rows connector.
Reasons to use Version B:
If you need a faster flow with less time between the last Excel table edits & the triggering of flow actions for those edits.
If you want to use this on an Excel table with more than 100,000 rows that is still below 100MB total content.
If you have a lower level Microsoft / Office365 license and only have a maximum pagination / Excel connector load of 5000.
If you have any trouble importing through the standard import method, see this post to import through a solution package.
Thanks for any feedback,
Please subscribe to my YouTube channel (https://youtube.com/@tylerkolota?si=uEGKko1U8D29CJ86).
And reach out on LinkedIn (https://www.linkedin.com/in/kolota/) if you want to hire me to consult or build more custom Microsoft solutions for you.
watch?v=85QAQ-tb1M8
For version B, I actually get a different error message.
For the 1st message the Solution package still imported & should be available to you in the options listed on that page.
The 2nd message, I’m guessing you tried to import the flow package as a solution package. You’d need to import it from the Power Automate legacy import, not the Power Apps solution import.
@takolota Thank you! I did figure both of those things out in the intervening hours. I'm a Power Automate newbie (as if you couldn't tell!).
@elizabethharms Actually the Threshold limit should be 5000.
Try importing via legacy with this version and it will work. https://powerusers.microsoft.com/jgvjg48436/attachments/jgvjg48436/MPA_Cookbook/2720/1/WhenanExcelro...
Thank you very much for your work. It benifited the community a lot.
There is a problem when trying to iterate over a dynamic chosen Excel file.
The scenario is that there is a folder A which all it's files will be deleted first.
After that an specific Excel file called "test.xlsx" will be copied into this folder.
Then, the Excel file will be read.
All that happens once a week with a scheduled automation.
The file contains about 12.000 rows. The flow works perfectly fine when selecting a static file.
First, I get the metada of the Excel file.
After that I, get it's properties to use the "file identifier" in the "List rows" action.
The "List rows" action works fine.
As you see, the second part - "List rows 2" - fails.
The cause of the error message is that the "item was not found" for the first batch.
The other two batches work fine.
I tried a lot of different approaches but can not figure out what the cause of the missing first batch for the dynamic chosen file is.
I hope, this post benefits everybody in the community.
Many thanks!
That looks like something unrelated to this build. Please make your post in the correct community forum: https://powerusers.microsoft.com/t5/Building-Flows/bd-p/BuildingFlows
Hi,
How we can trigger the flow when one specific column is modified in the excel ?
@DebarthaMitra You dont have to trigger it manually. As soon as any changes happens to the excel document, the flow will kick start
To narrow it down to only perform some actions if a specific column has changed you will likely need to use a Filter array on the Last table version to find the record with the matching primary key, then use a condition action to check if the column from the last table version is not equal to the column from the newest table version.
Here is something you can copy to clipboard, then go to a new action inside the Apply to each Modified loop & cntrl+V paste into the My clipboard tab so you can select it to insert it into the flow. It uses "YourColumnName" as the example column you are checking, so you will need to change the references in the Condition expressions from ?['YourColumnName'] to use the column you want to check, like for a column with the name Email, you would use ?['Email']
{"id":"0edc46d5-ea93-4a32-af26-07ee2509fa28","brandColor":"#8C3900","connectionReferences":{"shared_excelonlinebusiness":{"connection":{"id":"/new_sharedexcelonlinebusiness_84336"}},"shared_sharepointonline_1":{"connection":{"id":"/new_sharedsharepointonline_a8977"}}},"connectorDisplayName":"Control","icon":"data:image/svg+xml;base64,PHN2ZyB3aWR0aD0iMzIiIGhlaWdodD0iMzIiIHZlcnNpb249IjEuMSIgdmlld0JveD0iMCAwIDMyIDMyIiB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciPg0KIDxwYXRoIGQ9Im0wIDBoMzJ2MzJoLTMyeiIgZmlsbD0iIzhDMzkwMCIvPg0KIDxwYXRoIGQ9Im04IDEwaDE2djEyaC0xNnptMTUgMTF2LTEwaC0xNHYxMHptLTItOHY2aC0xMHYtNnptLTEgNXYtNGgtOHY0eiIgZmlsbD0iI2ZmZiIvPg0KPC9zdmc+DQo=","isTrigger":false,"operationName":"Copy_Scope","operationDefinition":{"type":"Scope","actions":{"Filter_array":{"type":"Query","inputs":{"from":"@outputs('Last_table_version')","where":"@equals(concat(item()?[outputs('Settings')?['KeyColumnName']], if(empty(item()?[outputs('Settings')?['KeyColumnName2']]), '', item()?[outputs('Settings')?['KeyColumnName2']])), concat(items('Apply_to_each_Modified')?[outputs('Settings')?['KeyColumnName']], if(empty(items('Apply_to_each_Modified')?[outputs('Settings')?['KeyColumnName2']]), '', items('Apply_to_each_Modified')?[outputs('Settings')?['KeyColumnName2']])))"},"runAfter":{}},"Condition":{"type":"If","expression":{"not":{"equals":["@first(body('Filter_array'))?['YourColumnName']","@items('Apply_to_each_Modified')?['YourColumnName']"]}},"actions":{},"runAfter":{"Filter_array":["Succeeded"]}}},"runAfter":{"Modified_referencing":["Succeeded"]}}}
Filter array
From
outputs('Last_table_version')
concat(item()?[outputs('Settings')?['KeyColumnName']], if(empty(item()?[outputs('Settings')?['KeyColumnName2']]), '', item()?[outputs('Settings')?['KeyColumnName2']]))
Is equal to
concat(items('Apply_to_each_Modified')?[outputs('Settings')?['KeyColumnName']], if(empty(items('Apply_to_each_Modified')?[outputs('Settings')?['KeyColumnName2']]), '', items('Apply_to_each_Modified')?[outputs('Settings')?['KeyColumnName2']]))
Condition
first(body('Filter_array'))?['YourColumnName']
Is equal to
items('Apply_to_each_Modified')?['YourColumnName']
If you use "Is not equal to" in the condition like in the picture, then you can put any actions you want to trigger on that column change in the green Yes side